Hi, I am Nandini Gunda, a Senior Data Engineer with 10+ years of experience designing and scaling data platforms across AWS, GCP, and Azure. I specialize in building batch and streaming pipelines, data modeling, governance, and delivering real-time insights with sub-minute latency. I love mentoring teams, guiding architectural decisions, and driving adoption of modern tools such as dbt, Delta Lake, and Apache Hudi. In my work, I partner with product and analytics teams to translate complex data into actionable insights, deliver reliable data infrastructure, and continuously improve data quality. When I am not coding, I enjoy exploring new cloud-native architectures and contributing to cross-functional initiatives that unlock business value.

Nandini Gunda

Hi, I am Nandini Gunda, a Senior Data Engineer with 10+ years of experience designing and scaling data platforms across AWS, GCP, and Azure. I specialize in building batch and streaming pipelines, data modeling, governance, and delivering real-time insights with sub-minute latency. I love mentoring teams, guiding architectural decisions, and driving adoption of modern tools such as dbt, Delta Lake, and Apache Hudi. In my work, I partner with product and analytics teams to translate complex data into actionable insights, deliver reliable data infrastructure, and continuously improve data quality. When I am not coding, I enjoy exploring new cloud-native architectures and contributing to cross-functional initiatives that unlock business value.

Available to hire

Hi, I am Nandini Gunda, a Senior Data Engineer with 10+ years of experience designing and scaling data platforms across AWS, GCP, and Azure. I specialize in building batch and streaming pipelines, data modeling, governance, and delivering real-time insights with sub-minute latency. I love mentoring teams, guiding architectural decisions, and driving adoption of modern tools such as dbt, Delta Lake, and Apache Hudi.

In my work, I partner with product and analytics teams to translate complex data into actionable insights, deliver reliable data infrastructure, and continuously improve data quality. When I am not coding, I enjoy exploring new cloud-native architectures and contributing to cross-functional initiatives that unlock business value.

See more

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Expert
Intermediate

Language

English
Fluent

Work Experience

Senior Data Engineer at Giant Eagle
October 1, 2025 - October 1, 2025
Led migration of 50+ legacy ETL jobs to AWS MWAA, reducing job failures by 45% and cutting operational overhead by over $100K annually. Built real-time streaming pipelines with Apache Kafka and Spark Structured Streaming to monitor SaaS metrics, achieving sub-minute latency. Optimized Spark batch pipelines processing 3TB+ daily, reducing transformation runtimes by 40% and lowering infra costs. Designed scalable data models in AWS Redshift for SaaS KPIs (ARR, CAC, CLTV) and implemented reusable ingestion/transformation frameworks. Introduced Great Expectations-based automated validation and anomaly detection, improving event-level data quality. Migrated S3 Parquet workflows to Apache Hudi for incremental upserts, boosting pipeline efficiency and maintainability. Implemented CI/CD with GitLab/Jenkins/Terraform/Helm and integrated Apache Atlas for lineage. Participated in cross-cloud PoCs comparing Spark performance on AWS EMR vs Azure Databricks and supported Delta Lake workflows.
Data Architect at Thomson Reuters
March 31, 2023 - March 31, 2023
Designed scalable, fault-tolerant data pipelines and ETL/ELT workflows for a SaaS environment, enabling batch and real-time analytics. Orchestrated Airflow DAGs with monitoring and alerting to ensure SLA adherence across environments. Delivered end-to-end streaming and batch solutions on Azure by ingesting high-volume Kafka data into Data Lake, Synapse Analytics, and Blob Storage. Optimized Azure Databricks pipelines processing 4TB+ daily, reducing latency by 35% and saving over $10K/month in compute costs. Built real-time event-stream processing with Kafka Streams and Spark Streaming on Azure Event Hubs, enabling near real-time user signals. Developed self-service datasets and Power BI dashboards for 100+ stakeholders, cutting insight delivery time by 50%. Led migration of 40+ batch jobs to stream-first architecture with zero data loss. Implemented CI/CD with Git, Terraform, Docker, and Azure DevOps. Mentored junior engineers in Python/Scala and Azure-native architecture.
Senior Data Engineer at Zendesk
September 30, 2021 - September 30, 2021
Designed and delivered scalable cloud-native pipelines on GCP (BigQuery, Cloud Storage, Composer), enabling global strategy and product teams to make timely, data-driven decisions. Built end-to-end ETL/ELT workflows with Cloud Composer, orchestrating Python and Spark jobs to process terabytes of data with high reliability. Architected real-time and batch data processing pipelines using Kafka, Spark Streaming, Scala, and Python, ingesting data into BigQuery and GCS for analytics. Refactored and optimized distributed Spark applications, boosting throughput by 45% and reducing Dataproc costs by $15K per quarter. Created self-service datasets and data libraries, reducing dependency on engineering teams. Integrated diverse SaaS sources (Salesforce, Stripe, Mixpanel) with internal logs into unified BigQuery datasets. Migrated legacy batch ETL to real-time event-driven architectures on GCP, significantly reducing latency and improving observability. Implemented Great Expectations for automate
Senior Data Engineer at Accenture India Pvt. Ltd.
February 28, 2019 - February 28, 2019
Designed and implemented robust database architectures with normalization up to 3NF, optimizing OLTP/OLAP workloads and improving SQL performance. Built predictive analytics and clustering models using Spark MLlib and Python, driving operational improvements. Streamlined Hadoop ecosystem operations by creating reusable frameworks and standardizing error handling and logging, enhancing debugging efficiency by 40%. Designed and deployed scalable ETL pipelines with Spark, Hive, and AWS-native services to populate an enterprise Data Lake. Implemented real-time monitoring with RBAC on Snowflake and BigQuery to protect sensitive data and ensure regulatory compliance. Led modernization of legacy Java systems with automated pipelines and peer reviews, delivering 25% performance improvements.
ETL Developer at Core Technologies
July 31, 2016 - July 31, 2016
Managed end-to-end data aspects of projects using Agile Scrum. Implemented ETL pipelines with Sqoop, Spark, and Hive for data ingestion into Hadoop Data Lake. Tuned SQL queries and optimized operations in AWS data environments. Developed Spark batch programs in PySpark/Scala, driving efficient testing and processing. Established CI/CD workflows with Jenkins and Maven, and used Apache NiFi for data movement automation across Hadoop systems.

Education

Bachelor of Science in Computer Science at Osmania University
January 1, 2010 - January 1, 2014

Qualifications

Add your qualifications or awards here.

Industry Experience

Software & Internet, Media & Entertainment, Professional Services

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Expert
Intermediate