Available to hire
I am a Sr. Data Engineer with extensive experience delivering data engineering solutions across banking, healthcare, finance, and insurance. I design secure, scalable data pipelines on cloud and on-prem, leveraging AWS, Spark, Hadoop, and modern data tooling to enable fast, reliable analytics and ML workflows.
I collaborate with data scientists, developers, and business partners to operationalize ML-ready datasets, govern data with HIPAA/PCI-DSS compliance, and automate ETL/ELT processes. I am passionate about building robust data platforms, enabling AI/ML initiatives, and driving measurable business impact.
Skills
See more
Work Experience
Sr. Data Engineer at Humana
August 1, 2025 - PresentDesign and manage large-scale ETL/ELT pipelines using Python, SQL, and Apache Spark to process claims, provider, enrollment, encounter, and care-management datasets. Cleaned and validated structured and semi-structured healthcare data with Pandas/NumPy while ensuring PHI handling aligns with HIPAA policies. Built RESTful APIs with FastAPI/Flask for secure data delivery to downstream systems. Supported production deployment of ML models (TensorFlow, PyTorch, Scikit-learn) for risk prediction, fraud detection, and member segmentation. Administered cloud-native infra on AWS, Azure, and GCP with storage, compute, networking, and IAM. Used Docker and Kubernetes for containerized data/ML services. Implemented CI/CD with Jenkins, GitHub Actions, and Azure DevOps. Built RAG/LLM-enabled apps using LangChain/LlamaIndex. Applied NLP to medical notes and communications and optimized queries across Snowflake/Redshift/Hadoop/Spark for cost efficiency. Enforced data quality with Great Expectations/De
Sr. Data Engineer at DaVita
April 1, 2023 - August 1, 2025Designed and maintained scalable data pipelines and ETL workflows using Python, SQL, and Apache Spark to support analytics, reporting, and AI/ML use cases. Processed structured and semi-structured data with Pandas/NumPy; built RESTful APIs via FastAPI/Flask. Partnered with data scientists to deploy TensorFlow/PyTorch/Scikit-learn models into production data pipelines. Managed cloud-native infra across AWS/Azure/GCP; containerized with Docker/Kubernetes; implemented CI/CD with Jenkins/GitHub Actions/Azure DevOps. Worked with LLMs and GenAI tools to develop intelligent data apps and RAG systems. Applied NLP and deep learning for document classification, entity extraction, and knowledge retrieval from unstructured sources. Improved query performance across Snowflake/Redshift/Hadoop/Spark. Enforced data quality through Great Expectations/Deequ/dbt tests and collaborated with cross-functional teams to translate business requirements into data solutions.
Senior Data Engineer at Mind Tree
March 1, 2022 - April 1, 2023Designed and managed ETL/ELT pipelines using Python, SQL, and Apache Spark to process healthcare data such as claims and care-management datasets. Used Pandas/NumPy for data cleaning and validation; built RESTful APIs with FastAPI/Flask. Collaborated with data scientists to deploy ML models and built cloud-native infrastructure across AWS/Azure/GCP. Used Docker/Kubernetes for containerized environments and implemented CI/CD with Jenkins/GitHub Actions/Azure DevOps. Worked with LangChain/LlamaIndex for intelligent data apps and RAG systems; applied NLP/deep learning to medical documents; optimized queries across Snowflake/Redshift/Hadoop. Ensured data quality with Great Expectations/Deequ/dbt; HIPAA/HITECH and GDPR compliance; partnered with BI/security/DevOps teams to deliver reliable data solutions.
Sr. Data Engineer at Liberty Mutual (Client)
March 1, 2021 - March 1, 2022Built high-throughput batch and streaming data pipelines using Apache Spark, Kafka, Flink, and Beam to process healthcare-related data from multiple systems. Developed ETL/ELT frameworks with AWS Glue, Airflow, dbt, and PySpark to ingest EHR/EMR data, claims logs, patient encounters, and device telemetry into S3 and Snowflake. Built real-time patient risk monitoring pipelines using Kafka Streams, AWS Lambda, and API integrations. Applied predictive analytics and ML models with Spark MLlib for risk, fraud, and anomaly detection. Implemented data quality monitoring with Great Expectations, AWS Deequ, dbt tests, and Bigeye; ensured HIPAA/HITECH and GDPR compliance; automated Terraform deployments for cloud infra. Led data governance and collaboration with care analytics and actuarial teams to ensure auditability and reliability.
Data Engineer at HSBC Electronic Data Processing Center
March 1, 2019 - March 1, 2021Led modernization of on-prem ETL by migrating DB2/iSeries pipelines to AWS Glue and Airflow, reducing processing time. Implemented multi-zone data lake on S3 with Lake Formation-based access control. Implemented CDC pipelines using DMS and Kafka Connect for near real-time data updates. Tuned Redshift performance with WLM, keys, and Spectrum integration. Built ML features with SageMaker Pipelines and Feature Store. Created dashboards with CloudWatch, Athena, and QuickSight for data freshness and SLA visibility. Enforced security through KMS encryption, TLS, IAM, and audit logging. Mentored juniors and standardized development workflows with Git. Ensured SOX and PCI-DSS compliance.
Data Engineer at Victor Express Line
March 1, 2016 - March 1, 2019Designed and implemented scalable ETL pipelines using AWS Glue and Airflow to process transactions, loans, mortgages, and customer data. Built real-time streaming pipelines with Kinesis and Kafka; developed ETL/ELT transformations with dbt and PySpark and applied star and snowflake schemas for risk analytics and fraud detection. Maintained legacy DB2/iSeries systems and batch programs; built S3 data lakes and tuned Redshift for analytics. Developed fraud detection pipelines using Spark MLlib, SageMaker, and Kinesis Data Analytics. Implemented data quality monitoring with Great Expectations, dbt tests, Lake Formation, and Apache Atlas to comply with PCI-DSS, SOX, Basel III, and GDPR. Automated infrastructure provisioning with Terraform and CI/CD using CodePipeline/GitHub Actions. Collaborated with risk, compliance, and business teams to deliver validated datasets for fraud analytics and regulatory reporting.
Senior Data Engineer at Sun Trade House
March 1, 2021 - March 1, 2022Designing and maintaining scalable data pipelines and ETL workflows to support analytics, reporting, and AI/ML use cases. Cleaning and transforming data from diverse sources and building RESTful APIs for downstream systems. Deploying ML models into production data pipelines and administering cloud-native infrastructure across AWS, Azure, and GCP. Implementing CI/CD pipelines to improve delivery speed. Working with LLMs and Generative AI tools to build intelligent data applications and RAG systems. Applying NLP and deep learning to extract insights from unstructured data and improve data quality with governance practices. Collaborating with cross-functional teams to deliver reliable data solutions.
Senior Data Engineer at MindTree
March 1, 2022 - April 1, 2023Designed and managed large-scale ETL/ELT pipelines using Python, SQL, and Apache Spark to process claims, provider, enrollment, encounter, and care-management datasets. Used Pandas/NumPy to clean, transform, and validate data while ensuring PHI handling aligns with HIPAA policies. Built and maintained RESTful APIs and backend data services with FastAPI/Flask; supported deployment of ML models into production pipelines. Administered cloud-native infrastructure across AWS/Azure/GCP; employed Docker/Kubernetes for containerization; implemented CI/CD with Jenkins, GitHub Actions, and Azure DevOps. Worked with LLM frameworks to build intelligent data applications and knowledge-retrieval tools; applied NLP and deep learning for healthcare data; optimized queries and jobs across Snowflake, Redshift, and Hadoop/Spark for cost efficiency and performance; ensured data quality with Great Expectations, Deequ, and dbt tests; collaborated with cross-functional teams to deliver reliable data solution
Sr. Data Engineer at Sun Trade House
March 1, 2021 - March 1, 2022Designed and maintained data pipelines and ETL workflows to support analytics and AI/ML use cases for Liberty Mutual client. Leveraged PySpark and Python to clean, transform, and analyze diverse data sources; built RESTful APIs for data delivery; partnered with data scientists to deploy ML models; managed cloud-native infrastructure across AWS/Azure/GCP; used Docker and Kubernetes to create scalable environments; streamlined CI/CD pipelines (Jenkins, GitHub Actions, Azure DevOps); explored LangChain/LlamaIndex for intelligent data applications and RAG systems; applied NLP for document classification and entity extraction; improved query performance across Snowflake/Redshift/Hadoop/Spark; ensured data quality with Great Expectations/Deequ/dbt tests; managed security/compliance; collaborated with care analytics, BI, security, and DevOps teams.
Education
Masters in Business Analytics at Ajeenkya D Y Patil University
January 11, 2030 - January 7, 2026Bachelors in Computer Science at Andhra University
January 11, 2030 - January 7, 2026Master's in Business Analytics at Ajeenkya D Y Patil University
January 11, 2030 - January 7, 2026Bachelor's in Computer Science at Andhra University
January 11, 2030 - January 7, 2026Master's in Business Analytics at Ajeenkya D Y Patil University
January 11, 2030 - January 7, 2026Bachelor's in Computer Science at Andhra University
January 11, 2030 - January 7, 2026Qualifications
Microsoft Certified: Azure Data Engineer
January 11, 2030 - January 7, 2026Google Certified Data Engineer
January 11, 2030 - January 7, 2026AWS Certified Data Engineer
January 11, 2030 - January 7, 2026Microsoft Certified: Azure Data Engineer
January 11, 2030 - January 7, 2026Google Certified: Data Engineer
January 11, 2030 - January 7, 2026AWS Certified: Data Engineer
January 11, 2030 - January 7, 2026Industry Experience
Healthcare, Financial Services, Professional Services, Software & Internet, Transportation & Logistics
Skills
See more
Hire a Full Stack Developer
We have the best full stack developer experts on Twine. Hire a full stack developer in Louisville today.