Hello, I’m Yeswanth Nidamanuri, a Senior Data Engineer with 7 years of enterprise experience delivering AI/ML, Generative AI, and MLOps solutions across Banking, Healthcare, Oil & Gas, and e-commerce. I design end-to-end AI pipelines across AWS, GCP, and Azure, integrating SageMaker, Vertex AI, Bedrock, Azure OpenAI, and Dataiku DSS to enable cross-cloud training, deployment, and governance. I build retrieval augmented generation (RAG) pipelines with LangChain, Pinecone, FAISS, Weaviate, and LlamaIndex to power semantic search and contextual reasoning in banking and healthcare. I apply agentic AI frameworks to orchestrate autonomous multi-agent systems for fraud investigations and compliance; secure LLM/API interactions with MCP; fine-tune LLMs with RLHF and LoRA for domain-specific needs; and drive production-grade MLOps with observability, governance, and reproducibility across complex enterprise environments.

Yeswanth Nidamanuri

Hello, I’m Yeswanth Nidamanuri, a Senior Data Engineer with 7 years of enterprise experience delivering AI/ML, Generative AI, and MLOps solutions across Banking, Healthcare, Oil & Gas, and e-commerce. I design end-to-end AI pipelines across AWS, GCP, and Azure, integrating SageMaker, Vertex AI, Bedrock, Azure OpenAI, and Dataiku DSS to enable cross-cloud training, deployment, and governance. I build retrieval augmented generation (RAG) pipelines with LangChain, Pinecone, FAISS, Weaviate, and LlamaIndex to power semantic search and contextual reasoning in banking and healthcare. I apply agentic AI frameworks to orchestrate autonomous multi-agent systems for fraud investigations and compliance; secure LLM/API interactions with MCP; fine-tune LLMs with RLHF and LoRA for domain-specific needs; and drive production-grade MLOps with observability, governance, and reproducibility across complex enterprise environments.

Available to hire

Hello, I’m Yeswanth Nidamanuri, a Senior Data Engineer with 7 years of enterprise experience delivering AI/ML, Generative AI, and MLOps solutions across Banking, Healthcare, Oil & Gas, and e-commerce. I design end-to-end AI pipelines across AWS, GCP, and Azure, integrating SageMaker, Vertex AI, Bedrock, Azure OpenAI, and Dataiku DSS to enable cross-cloud training, deployment, and governance.

I build retrieval augmented generation (RAG) pipelines with LangChain, Pinecone, FAISS, Weaviate, and LlamaIndex to power semantic search and contextual reasoning in banking and healthcare. I apply agentic AI frameworks to orchestrate autonomous multi-agent systems for fraud investigations and compliance; secure LLM/API interactions with MCP; fine-tune LLMs with RLHF and LoRA for domain-specific needs; and drive production-grade MLOps with observability, governance, and reproducibility across complex enterprise environments.

See more

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
See more

Language

English
Fluent

Work Experience

Senior Data Engineer with MLOps at Bank of America
October 1, 2023 - Present
Designed and deployed asynchronous ML inference pipelines using FastAPI, Celery, AWS Lambda, and ECS to deliver real-time fraud scoring and decisioning. Led hybrid ML workflows across AWS SageMaker and GCP Vertex AI, standardizing model governance across clouds and shortening training-to-deployment cycles by 25%. Fine-tuned fraud-detection LLMs (Bedrock + Hugging Face) with LoRA and RLHF, achieving 22% higher precision and reduced false positives in high-risk segments. Built knowledge graph–augmented reasoning for compliance, implemented RAG with LangChain + FAISS for rapid fraud-case retrieval, and designed LangGraph agentic workflows for autonomous multi-agent investigations. Optimized Elasticsearch with shard/replica strategies, implemented regex-based parsing and ontology alignment for fraud/compliance text analytics, and integrated MCP Protocol for secure LLM-API conversations. Implemented Pinecone-backed vector search, PSI drift detection across Dataiku, SageMaker, and Vertex A
Data Engineer II at Sempra
September 30, 2023 - October 2, 2025
Designed and deployed ML pipelines in Talend integrated with AWS SageMaker and GCP Vertex AI; automated retraining workflows with Talend, Lambda, and Dataflow to improve forecasting accuracy. Built Dataiku DSS workflows for energy demand predictions, integrated with S3 + BigQuery, and automated PSI-based drift detection. Developed containerized real-time energy forecasting microservices (FastAPI, Celery, EKS, GKE) and authored executable ML workflow plans bridging documentation with code execution. Implemented Core.ID federation with Azure AD for secure access, enforced identity-based restrictions, and built Logstash ingestion pipelines for high-volume sensor data. Tuned Elasticsearch indexing for real-time analytics, deployed Fivetran data replication across S3/Glue/Iceberg, documented data flows, and benchmarked pipelines for reliability. Prototyped agentic swarm forecasting in LangGraph/AutoGen, migrated Hadoop to Iceberg + Flink to reduce costs, and deployed Trino-based querying fo
Data Engineer at MedTourEasy
November 30, 2021 - October 2, 2025
Built HIPAA-compliant ETL pipelines using Talend + AWS Glue to ingest EHR, clinical, and claims data from APIs, Salesforce Health Cloud, and flat files into Redshift and BigQuery. Developed healthcare risk scoring models in Dataiku DSS, automated KPI dashboards, and integrated with Redshift/Glue pipelines. Applied NLP (Hugging Face transformers + GPT APIs) for summarization of patient engagement notes and secured EHR access via Core.ID. Implemented AI Ops monitoring with OTEL for ETL and inference, designed Excel/Tableau dashboards for risk score adoption and accuracy, and built claims anomaly detection workflows. Created step-by-step ML implementation plans for regulatory readiness, benchmarked anomaly detection models, and fine-tuned transformers with RLHF on HIPAA datasets. Designed Logstash parsing for clinical data, Kibana dashboards, onboarded healthcare data sources with Fivetran, and applied metadata lineage. Built RAG pipelines with FAISS embeddings for contextual patient hist
Data Analyst at Futures First
August 31, 2020 - October 2, 2025
Built HIPAA-compliant ETL pipelines using Talend + Glue to ingest EHR, clinical, and claims data from APIs, Salesforce Health Cloud, and flat files into Redshift and BigQuery. Developed healthcare risk scoring models in Dataiku DSS, automating KPI dashboards and integrating with Redshift/Glue pipelines for patient retention tracking. Applied NLP models (Hugging Face Transformers + GPT APIs) to summarize patient engagement notes, improving medical staff productivity. Implemented Core.ID-based authentication and role management for financial risk dashboards, with identity-aware logging to support compliance audits. Applied OTEL observability for healthcare ETL/inference pipelines, built Kibana visuals for trading datasets, configured Beats for streaming logs, and optimized indexing. Built Fivetran ingestion workflows for replicating financial datasets into Snowflake/Redshift, monitored pipeline health, and designed Excel/Tableau dashboards to track risk and adoption. Engineered claims an
Senior Data Engineer with MLOps at Bank of America
October 1, 2023 - Present
Designed and deployed asynchronous ML inference pipelines using FastAPI, Celery, AWS Lambda, and ECS to enable real-time fraud scoring with reduced latency. Implemented hybrid training and deployment workflows across AWS SageMaker and GCP Vertex AI, standardizing governance across clouds. Fine-tuned fraud-detection LLMs with LoRA and RLHF, achieving higher precision and fewer false positives. Built knowledge-graph augmented reasoning pipelines for compliance reviews and retrieval-augmented generation (RAG) systems with LangChain and vector stores. Implemented MCP Protocol for secure LLM-API communications. Created Pinecone-backed vector search APIs for fast semantic queries, and integrated PSI drift detection to trigger retraining. Migrated Hadoop pipelines to Iceberg + Trino for speed and governance. Configured Elasticsearch clusters, Logstash, Beats, and Kibana dashboards for fraud and compliance insights. Integrated Core.ID for identity management, and established OTEL observability
Data Engineer II at Sempra
September 30, 2023 - October 2, 2025
Designed and deployed ML pipelines in Talend integrated with AWS SageMaker and GCP Vertex AI, reducing training time and enabling automated retraining. Built Dataiku DSS workflows for energy forecasting with S3 and BigQuery, and integrated PSI drift detection to improve forecast accuracy. Developed containerized inference microservices (FastAPI, Celery, EKS, GKE) for real-time energy forecasting. Implemented Core.ID federation with Azure AD to secure access and enforced identity-based restrictions in ML pipelines. Built Logstash ingestion pipelines for high-volume sensor data; tuned Elasticsearch for performance at scale; created Trino-based query engines on Iceberg. Documented data flows and decisions for reproducibility. Prototyped agentic swarm forecasting in LangGraph and AutoGen. Migrated Hadoop to Iceberg + Flink, reducing compute costs and improving scalability. Integrated SHAP + LIME explainability into forecasting models and built dashboards in Tableau/Excel. Established CI/CD
Data Engineer at MedTourEasy
November 30, 2021 - October 2, 2025
Built HIPAA-compliant ETL pipelines using Talend + AWS Glue to ingest EHR, clinical, and claims data into Redshift and BigQuery. Developed healthcare risk scoring models in Dataiku DSS and automated KPI dashboards linked to Redshift/Glue pipelines. Applied NLP with Hugging Face transformers and GPT APIs to summarize patient engagement notes. Implemented Core.ID federation for HIPAA-compliant access and identity-based restrictions for PII/PHI. Implemented AI Ops with OTEL observability across ETL and inference pipelines. Created dashboards in Excel/Tableau for adoption and accuracy, developed claims anomaly detection to reduce fraud leakage, and documented end-to-end ML implementation plans. Fine-tuned transformers with RLHF on HIPAA data for clinical summarization. Built Logstash pipelines for clinical/claims data and Kibana dashboards; onboarded new healthcare data sources with Fivetran into Glue/Redshift pipelines. Implemented metadata cataloging and lineage tracking; built RAG pipel
Data Analyst at Futures First
August 31, 2020 - October 2, 2025
Configured Core.ID-based authentication and role management for financial risk dashboards; implemented identity-aware logging and audit trails for compliance. Built Kibana visualizations for trading datasets, and configured Beats agents to stream logs into Elasticsearch. Optimized indexing and queries for large-scale trading data; built Fivetran ingestion workflows for replicating datasets into Snowflake/Redshift. Monitored pipeline health for latency and data quality; designed Excel/Tableau dashboards to track adoption and accuracy of risk models. Developed structured ML experiment plans, validated outputs against specifications, and benchmarked reproducibility. Implemented RAG pipelines with FAISS embeddings for contextual trading insights; packaged Python APIs into Helm-deployed Kubernetes services; applied data masking/encryption with AWS.
Senior Data Engineer with MLOps at Bank of America
October 1, 2023 - Present
Designed and deployed asynchronous ML inference pipelines using FastAPI, Celery, AWS Lambda, and ECS; built hybrid ML workflows across AWS SageMaker and GCP Vertex AI, standardizing model governance across clouds; fine-tuned fraud-detection LLMs with LoRA and RLHF; implemented retrieval augmented generation (RAG) with LangChain + FAISS for real-time fraud case retrieval; built LangGraph-based multi-agent workflows to automate fraud investigations; configured Elasticsearch with optimized shard/replica layouts; implemented MCP for secure LLM-API interactions; developed Pinecone-backed vector search for semantic fraud queries; established PSI drift detection and synthetic data generation; migrated Hadoop pipelines to Iceberg + Trino to improve performance and governance; deployed CI/CD with Helm, Jenkins, and GitHub Actions; introduced OTEL observability across ML microservices; supported cross-functional teams with Jira/Confluence integration; delivered explainability dashboards (SHAP/LI
Data Engineer II at Sempra
September 30, 2023 - October 15, 2025
Designed and deployed ML pipelines using Talend integrated with AWS SageMaker and GCP Vertex AI; automated retraining workflows with Talend, Lambda, and Dataflow; built Dataiku DSS workflows for energy demand forecasting; developed containerized real-time inference microservices (FastAPI, Celery, EKS/GKE); implemented Core.ID federation with Azure AD for secure access; created Logstash ingestion and Kibana dashboards to monitor demand anomalies; tuned Elasticsearch performance at scale; established real-time data replication with Fivetran (S3/Glue/Iceberg); migrated Hadoop pipelines to Iceberg + Flink to reduce compute costs; implemented CI/CD with GitHub Actions, Terraform, and Helm; introduced OTEL + Prometheus/Grafana monitoring; prototyped agentic swarm forecasting with LangGraph/AutoGen; integrated SHAP + LIME explanations into BigQuery/Redshift forecasting; delivered executive dashboards; ensured IAM/KMS-based data protection.
Data Engineer at MedTourEasy
November 30, 2021 - October 15, 2025
HIPAA-compliant ETL pipelines using Talend + AWS Glue ingesting EHR, clinical, and claims data into Redshift and BigQuery; developed Dataiku DSS risk scoring models for patient retention; applied NLP (Hugging Face transformers) to summarize clinical notes; integrated Core.ID for HIPAA-compliant data access; engineered tokenization strategies for medical terminology; built AI Ops monitoring with OTEL; created dashboards in Excel and Tableau; implemented RAG pipelines with FAISS; packaged Python APIs into Helm-deployed Kubernetes services; applied data masking and encryption for PII/PHI; built chatbots and dashboards with Flask; implemented event-driven ingestion (S3, SNS, Step Functions) for near real-time telemetry; designed fraud detection ML models for claims using structured and unstructured data.
Data Analyst at Futures First
August 31, 2020 - October 15, 2025
Built HIPAA-compliant ETL pipelines using Talend + Glue to ingest EHR, clinical, and claims data from APIs, Salesforce Health Cloud, and flat files into Redshift and BigQuery; developed healthcare risk scoring models in Dataiku DSS; applied NLP models (Hugging Face Transformers + GPT APIs) to summarize patient engagement notes; configured Core.ID-based authentication for financial risk dashboards; implemented identity-aware logging and audit trails to support compliance; built Kibana visualizations for trading datasets; configured Beats agents to stream log data; tuned indexing and queries for large-scale data; built Fivetran-based ingestion for Snowflake and Redshift; monitored pipeline health; designed dashboards to track adoption and accuracy of risk models; developed RAG pipelines with FAISS embeddings for contextual trading analyses; packaged Python APIs into Helm-deployed Kubernetes services; applied event-driven ingestion (S3 triggers, SNS, Step Functions) for telemetry and lab

Education

Add your educational history here.

Qualifications

Add your qualifications or awards here.

Industry Experience

Healthcare, Financial Services, Energy & Utilities, Software & Internet, Professional Services