I'm a Senior Data Engineer with a passion for building secure, scalable ETL/ELT pipelines across AWS, Azure, and Snowflake. I design data models and pipelines using AWS Glue, Glue Studio, PySpark, and Python, ingesting structured and semi-structured data into S3 and processing them through automated cloud workflows. I enjoy turning complex data into reliable, analytics-ready datasets for reporting, data science, and AI initiatives, while enforcing governance, data quality, and lineage. I thrive collaborating with analysts, data scientists, and business stakeholders to translate needs into robust data architectures. My experience spans healthcare, retail, and financial analytics, with HIPAA, GDPR, and SOC2 compliance in mind. I'm comfortable optimizing Spark jobs, building secure Snowflake models with RBAC and secure views, and orchestrating end-to-end pipelines with Airflow, GitLab CI/CD, and containerized deployments.

Uma Devi

I'm a Senior Data Engineer with a passion for building secure, scalable ETL/ELT pipelines across AWS, Azure, and Snowflake. I design data models and pipelines using AWS Glue, Glue Studio, PySpark, and Python, ingesting structured and semi-structured data into S3 and processing them through automated cloud workflows. I enjoy turning complex data into reliable, analytics-ready datasets for reporting, data science, and AI initiatives, while enforcing governance, data quality, and lineage. I thrive collaborating with analysts, data scientists, and business stakeholders to translate needs into robust data architectures. My experience spans healthcare, retail, and financial analytics, with HIPAA, GDPR, and SOC2 compliance in mind. I'm comfortable optimizing Spark jobs, building secure Snowflake models with RBAC and secure views, and orchestrating end-to-end pipelines with Airflow, GitLab CI/CD, and containerized deployments.

Available to hire

I’m a Senior Data Engineer with a passion for building secure, scalable ETL/ELT pipelines across AWS, Azure, and Snowflake. I design data models and pipelines using AWS Glue, Glue Studio, PySpark, and Python, ingesting structured and semi-structured data into S3 and processing them through automated cloud workflows. I enjoy turning complex data into reliable, analytics-ready datasets for reporting, data science, and AI initiatives, while enforcing governance, data quality, and lineage.

I thrive collaborating with analysts, data scientists, and business stakeholders to translate needs into robust data architectures. My experience spans healthcare, retail, and financial analytics, with HIPAA, GDPR, and SOC2 compliance in mind. I’m comfortable optimizing Spark jobs, building secure Snowflake models with RBAC and secure views, and orchestrating end-to-end pipelines with Airflow, GitLab CI/CD, and containerized deployments.

See more

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Intermediate
Intermediate

Work Experience

Data Warehouse Developer at UW-Madison School of Medicine & Public Health
March 1, 2025 - Present
Led the design of conceptual, logical, and physical data models using DBSchema to align with enterprise standards, ensuring scalability and analytics-readiness. Defined fact table grains, surrogate keys, and dimension relationships to preserve history and optimize Snowflake query performance. Authored transformation rules and metadata documentation for full transparency across Informatica IICS ETL workflows and reporting layers. Evaluated source entities to identify high-value datasets for Bronze layer ingestion, establishing minimal transformation guidelines to preserve lineage while meeting governance and compliance mandates. Developed Python automation for metadata validation, schema checks, and key structure enforcement. Built scalable ETL pipelines ingesting structured and semi-structured healthcare data into Snowflake, with automated quality checks and performance tuning. Designed incremental load strategies in Informatica IICS to minimize latency. Created durable Snowflake data
Senior Data Engineer at AbbVie
September 1, 2024 - February 28, 2025
Designed and developed large-scale ETL pipelines on AWS Glue (PySpark), Glue Studio, and Glue Catalog, integrating clinical, commercial, and R&D datasets. Built optimized PySpark transformation scripts inside Glue Jobs using partition tuning, pushdown filters, broadcast joins, and job bookmarks for incremental processing. Ingested structured and semi-structured datasets into Amazon S3, applying schema enforcement, automated validation rules, and metadata tagging. Developed analytical datasets in Amazon Redshift and Athena, optimizing table design, compression, and query execution for high-volume scientific workloads. Built event-driven workflows using AWS Lambda, Step Functions, and CloudWatch for orchestration, alerting, and automated error recovery. Implemented end-to-end data quality checks, lineage tracking, and audit controls across all AWS pipelines to maintain compliance with healthcare regulations. Designed secure and governed models in Snowflake using secure views, row-level s
Data Analyst/Engineer at Ace Hardware Corporation
October 1, 2019 - December 31, 2022
Developed secure retail ETL workflows in Informatica PowerCenter and Azure Data Factory (ADF) for sales, store, and inventory datasets with strong credential and access governance. Designed Python and SQL transformation logic for SKU-level and warehouse-level analytics used in forecasting, merchandising, and replenishment. Ingested semi-structured and unstructured data into AWS S3 and Azure, applying schema enforcement and automated validation rules. Maintained Spark Streaming jobs for IoT, logistics, and clickstream pipelines using PySpark, applying schema-first design and secure checkpointing. Prepared clean Power BI models for store performance dashboards, category analysis, and supply chain KPIs with Row-Level Security. Developed curated retail datasets that became the default source for merchandising, inventory, and supply-chain analytics. Improved nightly retail loads by optimizing PowerCenter mappings, SQL queries, and ADF runtimes for time-sensitive operational reporting. Imple
Data Analyst at Thomson Reuters
April 1, 2017 - August 31, 2019
Designed and developed SSIS ETL packages to extract, transform, and load financial and regulatory data into SQL Server data warehouses and marts. Implemented incremental load logic, error handling, and retry mechanisms to improve job reliability and meet reporting SLAs. Built complex data transformations using SSIS expressions, conditional splits, and lookups to apply business and compliance rules. Migrated high-volume reporting workloads from on-prem SQL systems to Amazon Redshift, reducing query time and improving scalability. Assisted in pilot migration of analytical workloads to Google BigQuery, optimizing SQL queries and validating accuracy across both platforms. Performed data profiling, data quality checks, and validation of financial data to ensure consistency across reporting layers. Automated daily ETL job scheduling and monitoring using SQL Server Agent. Created and maintained Tableau Server dashboards with row-level security and user filters. Collaborated with business anal
Data Analyst at Merck Healthcare
June 1, 2015 - March 31, 2017
Collected and prepared sales, production, and distribution data from SQL Server, Excel, and flat files to support business and compliance reporting. Wrote and optimized SQL queries, views, and stored procedures to analyze product sales trends and inventory performance. Assisted in developing ETL workflows using Informatica PowerCenter and SQL scripts to load ERP data into centralized reporting databases. Created Excel-based dashboards using pivot tables, charts, and macros to track sales targets, stock movement, and production output. Supported the rollout of Power BI Desktop for regional analytics, connecting on-prem SQL databases to build visual reports for sales and marketing teams. Performed data validation and reconciliation between SAP extracts and SQL datasets to ensure consistency in business and financial reports. Helped migrate legacy data from Access databases and flat files into SQL Server for better traceability and unified analytics. Worked with senior data engineers on p

Education

BTech in Computer Science and Engineering at Jawaharlal Nehru Technological University, Hyderabad, TS, India
June 1, 2011 - May 31, 2015

Qualifications

Add your qualifications or awards here.

Industry Experience

Healthcare, Retail, Financial Services

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Intermediate
Intermediate