Dynamic Data Engineer with hands-on experience designing and implementing scalable big data solutions using Apache Spark and MapReduce. Proficient in Python, Scala, and Java, with strong experience building cloud-native data platforms on AWS using services such as S3, AWS Glue, EMR, Lambda, Redshift, Athena, and Kinesis. Proven ability to process large-scale structured and semi-structured datasets, optimize ETL and ELT workflows, and design high-performance OLAP and OLTP systems. Experienced in applying solid data modeling techniques, including dimensional modeling, star schemas, and slowly changing dimensions, to support analytics and reporting. Strong focus on data quality, validation, and governance, implementing checks for completeness, consistency, and accuracy across pipelines. Well-versed in delivering data solutions across healthcare, finance, and telecom domains, with strong problem-solving and stakeholder communication skills to translate complex data architectures into actionable insights.

Vindhya Reddy

Dynamic Data Engineer with hands-on experience designing and implementing scalable big data solutions using Apache Spark and MapReduce. Proficient in Python, Scala, and Java, with strong experience building cloud-native data platforms on AWS using services such as S3, AWS Glue, EMR, Lambda, Redshift, Athena, and Kinesis. Proven ability to process large-scale structured and semi-structured datasets, optimize ETL and ELT workflows, and design high-performance OLAP and OLTP systems. Experienced in applying solid data modeling techniques, including dimensional modeling, star schemas, and slowly changing dimensions, to support analytics and reporting. Strong focus on data quality, validation, and governance, implementing checks for completeness, consistency, and accuracy across pipelines. Well-versed in delivering data solutions across healthcare, finance, and telecom domains, with strong problem-solving and stakeholder communication skills to translate complex data architectures into actionable insights.

Available to hire

Dynamic Data Engineer with hands-on experience designing and implementing scalable big data solutions using Apache Spark and MapReduce. Proficient in Python, Scala, and Java, with strong experience building cloud-native data platforms on AWS using services such as S3, AWS Glue, EMR, Lambda, Redshift, Athena, and Kinesis. Proven ability to process large-scale structured and semi-structured datasets, optimize ETL and ELT workflows, and design high-performance OLAP and OLTP systems. Experienced in applying solid data modeling techniques, including dimensional modeling, star schemas, and slowly changing dimensions, to support analytics and reporting. Strong focus on data quality, validation, and governance, implementing checks for completeness, consistency, and accuracy across pipelines. Well-versed in delivering data solutions across healthcare, finance, and telecom domains, with strong problem-solving and stakeholder communication skills to translate complex data architectures into actionable insights.

See more

Work Experience

Data Engineer at AstraZeneca
October 1, 2023 - Present
Optimized PostgreSQL through SQL query refactoring, tuning by reducing the query retrieval time from 90 secs to 20 secs, reducing table scans for over 20 million records. Managed Fivetran pipelines to automate integration of 10+ data sources, reducing manual data handling by 40%. Orchestrated ETL/ELT jobs: daily, weekly, and monthly to process data from disparate data sources from S3 to data warehouse using various sensors and operators in Airflow.
Technology Consulting Analyst- Data at Accenture Ltd
May 1, 2022 - August 1, 2023
Built audit trail and data lineage reports using dbt artifacts + Snowflake metadata for regulatory inspections, improving data traceability coverage by 100%. Streamlined CI/CD pipelines using Github for version controlling and Jenkins while following DevOps best practices.
Associate Data Engineer at Virtusa Consulting Services
June 1, 2018 - December 1, 2019
Designed and implemented a scalable Hadoop-based data pipeline to process massive volumes of telecom billing and transactional data, improving data processing efficiency by 45%. Utilized Apache Nifi data ingestion pipelines to efficiently ingest 50+ tables into Hive, ensuring seamless data integration and transformation.

Education

Master of Business Administration at Symbiosis International University
January 1, 2020 - January 1, 2022
Bachelor of Technology in Computer Science at Jawaharlal Nehru Technological University
January 11, 2030 - January 7, 2026

Qualifications

AWS Certified Data Engineer Associate
September 1, 2024 - January 7, 2026
Microsoft Azure Data Fundamentals
July 1, 2022 - January 7, 2026
Entry Certificate on Business Analyst
June 1, 2022 - January 7, 2026
HDP Certified Spark Developer
June 1, 2018 - January 7, 2026
Oracle Certified Associate- Java SE8
June 1, 2017 - January 7, 2026

Industry Experience

Healthcare, Life Sciences

Hire a Data Analyst

We have the best data analyst experts on Twine. Hire a data analyst in Wilmington today.