Banner Image

All Services

Programming & Development Database Design & Administration

Principal Data Engineer | AWS/GCP/Azure

$20/hr Starting at $100

I help companies build scalable, production-grade data pipelines that handle millions of events reliably from ingestion to analytics, using Kafka, Airflow, Databricks, and cloud platforms. With 11 years of experience in designing scalable ETL/ELT pipelines, building data platforms, and optimizing high-performance data warehouses, I specialize in technologies such as Kafka, NiFi, Airflow, Databricks, and cloud platforms (AWS, GCP, Azure). I deliver production-grade solutions that enable real-time data processing, business intelligence, and machine learning. 

🔹 What I Can Do for You: 

Database Architecture & Optimization → Design and tune high-performance data models, cutting query time from hours to minutes. 

Design and build scalable data pipelines (batch + streaming) using Python, SQL, Apache Kafka, Apache Spark, Airflow, and cloud platforms (AWS, GCP, Azure). 

Develop data lakes, data warehouses, and lakehouse architectures that support high-performance analytics and real-time reporting. Implement data validation, monitoring and CI/CD automation to ensure reliability and maintainability. 

Optimize data workflows for cost, performance, and scalability while adhering to best practices in architecture and design. 

Collaborate closely with analytics and ML teams to deliver feature-ready datasets for business intelligence and machine learning workflows. 

🔹 Tech Stack: 

Programming & Tools: Python (Pandas, Flask, Django, FastAPI, REST APIs), SQL, Bash Scripting, Spark 

Cloud Platforms: AWS (S3, Redshift, RDS, Lambda, EMR, Athena), GCP (BigQuery, GKE, Cloud Storage), Azure 

Data Engineering: Apache Kafka, Apache NiFi, Apache Airflow, Databricks, Delta Lake, Hive, ELK Stack (Elasticsearch, Logstash, Kibana), FluentD, Informatica (PowerCenter) 

DevOps & CI/CD: Docker, Kubernetes, Linux, AWS CloudFormation, Git, GitHub Actions, GitLab CI/CD 

BI & Visualization: Power BI, Kibana, Amazon QuickSight 

Others: Data Warehousing, ETL / ELT Pipelines, Real-Time Streaming, Data Integration, Data Pipelines, Microservices Architecture, Star & Snowflake Schema, Data Validation, Monitoring & Lineage, Logging & Alerting Pipelines 

🔹 Why Clients Hire Me:

Enterprise-grade experience (GSK) 

Clear communication & documentation

Ownership mindset, I treat your data as critical infrastructure 

Focus on business outcomes, not just tools

📩 If you need a Data Engineer who can own your data workflows end-to-end, let’s talk.

About

$20/hr Ongoing

Download Resume

I help companies build scalable, production-grade data pipelines that handle millions of events reliably from ingestion to analytics, using Kafka, Airflow, Databricks, and cloud platforms. With 11 years of experience in designing scalable ETL/ELT pipelines, building data platforms, and optimizing high-performance data warehouses, I specialize in technologies such as Kafka, NiFi, Airflow, Databricks, and cloud platforms (AWS, GCP, Azure). I deliver production-grade solutions that enable real-time data processing, business intelligence, and machine learning. 

🔹 What I Can Do for You: 

Database Architecture & Optimization → Design and tune high-performance data models, cutting query time from hours to minutes. 

Design and build scalable data pipelines (batch + streaming) using Python, SQL, Apache Kafka, Apache Spark, Airflow, and cloud platforms (AWS, GCP, Azure). 

Develop data lakes, data warehouses, and lakehouse architectures that support high-performance analytics and real-time reporting. Implement data validation, monitoring and CI/CD automation to ensure reliability and maintainability. 

Optimize data workflows for cost, performance, and scalability while adhering to best practices in architecture and design. 

Collaborate closely with analytics and ML teams to deliver feature-ready datasets for business intelligence and machine learning workflows. 

🔹 Tech Stack: 

Programming & Tools: Python (Pandas, Flask, Django, FastAPI, REST APIs), SQL, Bash Scripting, Spark 

Cloud Platforms: AWS (S3, Redshift, RDS, Lambda, EMR, Athena), GCP (BigQuery, GKE, Cloud Storage), Azure 

Data Engineering: Apache Kafka, Apache NiFi, Apache Airflow, Databricks, Delta Lake, Hive, ELK Stack (Elasticsearch, Logstash, Kibana), FluentD, Informatica (PowerCenter) 

DevOps & CI/CD: Docker, Kubernetes, Linux, AWS CloudFormation, Git, GitHub Actions, GitLab CI/CD 

BI & Visualization: Power BI, Kibana, Amazon QuickSight 

Others: Data Warehousing, ETL / ELT Pipelines, Real-Time Streaming, Data Integration, Data Pipelines, Microservices Architecture, Star & Snowflake Schema, Data Validation, Monitoring & Lineage, Logging & Alerting Pipelines 

🔹 Why Clients Hire Me:

Enterprise-grade experience (GSK) 

Clear communication & documentation

Ownership mindset, I treat your data as critical infrastructure 

Focus on business outcomes, not just tools

📩 If you need a Data Engineer who can own your data workflows end-to-end, let’s talk.

Skills & Expertise

ApacheCommunication SkillsData ManagementData ProcessingData WarehouseDatabase AdministrationDatabase DesignDatabase DevelopmentDB2DesignEngineeringGitHubGraph DatabaseMicrosoft SQL ServerMongoDBMySQLOraclePL/SQLPostgreSQLPythonQuery OptimizationSQLSQLiteStored Procedures

0 Reviews

This Freelancer has not received any feedback.