I'm a data engineer with 10 years of experience building scalable data platforms across cloud, big data, and enterprise ecosystems. I specialize in end-to-end ETL/ELT pipelines using Python, SQL, Spark, Hadoop, and Kafka, enabling reliable data ingestion, processing and analytics at scale. I enjoy collaborating with cross-functional teams, implementing governance, and delivering business value through data lakes, data warehouses, and ML-ready architectures on AWS, Azure, and GCP. I thrive in Agile environments, mentoring teammates, and continuously optimizing performance, cost, and reliability.

Sai Eswar Deepak Pamarthi

I'm a data engineer with 10 years of experience building scalable data platforms across cloud, big data, and enterprise ecosystems. I specialize in end-to-end ETL/ELT pipelines using Python, SQL, Spark, Hadoop, and Kafka, enabling reliable data ingestion, processing and analytics at scale. I enjoy collaborating with cross-functional teams, implementing governance, and delivering business value through data lakes, data warehouses, and ML-ready architectures on AWS, Azure, and GCP. I thrive in Agile environments, mentoring teammates, and continuously optimizing performance, cost, and reliability.

Available to hire

I’m a data engineer with 10 years of experience building scalable data platforms across cloud, big data, and enterprise ecosystems. I specialize in end-to-end ETL/ELT pipelines using Python, SQL, Spark, Hadoop, and Kafka, enabling reliable data ingestion, processing and analytics at scale.

I enjoy collaborating with cross-functional teams, implementing governance, and delivering business value through data lakes, data warehouses, and ML-ready architectures on AWS, Azure, and GCP. I thrive in Agile environments, mentoring teammates, and continuously optimizing performance, cost, and reliability.

See more

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert

Language

English
Fluent

Work Experience

Sr. Data Engineer at Cigna
November 1, 2023 - Present
Led ETL workflows with AWS Glue to process large datasets reliably. Built scalable Python-based transformations with Pandas/NumPy; developed batch and streaming pipelines using PySpark/Spark SQL. Implemented schema transformations to normalize datasets for enterprise analytics and applied Kimball dimensional modeling and Data Vault 2.0 concepts for warehouse design. Delivered business-ready data marts and dashboards; established governance with IAM, tagging, and metadata management via AWS Glue Data Catalog and Databricks Unity Catalog. Automated event-driven processing with AWS Lambda and Step Functions; integrated data from Hadoop/HDFS/Hive environments and built RESTful connectors for downstream systems.
Data Engineer at FM Global
September 1, 2021 - October 1, 2023
Built streaming and batch data ingestion pipelines using Kafka (Streams/Connect) and Airflow; automated workflows reducing manual effort. Implemented distributed processing with Apache Spark on Azure Databricks, and designed scalable cloud-based architectures with ADLS and Snowflake. Developed Power BI dashboards and collaborated with ML teams using Azure ML and Kubeflow. Authored complex SQL queries for lakehouse/warehouse environments, established data governance with metadata catalogs, and integrated legacy systems (Hadoop, Hive, HDFS) for migration.
Data Engineer at FuGenX Technologies, India
January 1, 2019 - July 1, 2021
Designed scalable ETL pipelines in GCP using Informatica and Matillion; built ingestion frameworks for NoSQL data (MongoDB). Implemented low-latency data delivery with Google Pub/Sub; developed ML-ready pipelines in Python and Dataflow; built Spark jobs on Dataproc and integrated Cloud Storage with Snowflake. Created secure RESTful APIs, tuned Spark jobs for performance, and delivered executive dashboards in Tableau/BigQuery. Automated CI/CD with Jenkins; implemented metadata-driven ingestion and governance practices.
Data Engineer at FuGenX Technologies, India
July 1, 2016 - December 31, 2018
Built end-to-end data ingestion and transformation pipelines with Apache Beam; migrated large relational datasets to HDFS via Sqoop; configured AWS EC2 clusters for high-throughput processing. Developed batch jobs with Talend and MapReduce, created Hive tables with partitioning/bucketing, and built a secure AWS S3 data lake. Implemented schema validation, data governance, and lineage tracking; automated ETL workflows with Lambda, and supported Agile CI/CD practices.

Education

Bachelor of Technology in Electronics & Communication Engineering at Koneru Lakshmaiah University (KLU)
January 11, 2030 - December 8, 2025

Qualifications

Add your qualifications or awards here.

Industry Experience

Software & Internet, Professional Services, Healthcare, Education, Media & Entertainment