I am an Artificial Intelligence Engineer with over 7 years of experience, specializing as a Senior Data Engineer. I have a proven track record of accelerating data processing by 50%, reducing costs by 40%, and achieving 99.9% pipeline uptime using technologies such as Spark MLlib, PySpark, and Kafka across leading cloud platforms including AWS, GCP, and Azure. My expertise lies in designing high-performance ETL workflows, optimizing big data pipelines, and building advanced analytical models to support enterprise-scale decision-making. Throughout my career, I have led the development of GPT-4-powered financial assistant applications, real-time fraud detection APIs, and scalable ML pipelines that process millions of transactions daily. I am skilled in automating regulatory reporting, establishing robust CI/CD pipelines, and developing innovative AI solutions that boost recommendation precision and improve operational efficiency. I continuously strive to deliver impactful AI and data engineering products that enhance user experience and drive business value.

Sushma Reddy

I am an Artificial Intelligence Engineer with over 7 years of experience, specializing as a Senior Data Engineer. I have a proven track record of accelerating data processing by 50%, reducing costs by 40%, and achieving 99.9% pipeline uptime using technologies such as Spark MLlib, PySpark, and Kafka across leading cloud platforms including AWS, GCP, and Azure. My expertise lies in designing high-performance ETL workflows, optimizing big data pipelines, and building advanced analytical models to support enterprise-scale decision-making. Throughout my career, I have led the development of GPT-4-powered financial assistant applications, real-time fraud detection APIs, and scalable ML pipelines that process millions of transactions daily. I am skilled in automating regulatory reporting, establishing robust CI/CD pipelines, and developing innovative AI solutions that boost recommendation precision and improve operational efficiency. I continuously strive to deliver impactful AI and data engineering products that enhance user experience and drive business value.

Available to hire

I am an Artificial Intelligence Engineer with over 7 years of experience, specializing as a Senior Data Engineer. I have a proven track record of accelerating data processing by 50%, reducing costs by 40%, and achieving 99.9% pipeline uptime using technologies such as Spark MLlib, PySpark, and Kafka across leading cloud platforms including AWS, GCP, and Azure. My expertise lies in designing high-performance ETL workflows, optimizing big data pipelines, and building advanced analytical models to support enterprise-scale decision-making.

Throughout my career, I have led the development of GPT-4-powered financial assistant applications, real-time fraud detection APIs, and scalable ML pipelines that process millions of transactions daily. I am skilled in automating regulatory reporting, establishing robust CI/CD pipelines, and developing innovative AI solutions that boost recommendation precision and improve operational efficiency. I continuously strive to deliver impactful AI and data engineering products that enhance user experience and drive business value.

See more

Work Experience

Senior AI/ML Engineer at Truist Bank
June 1, 2024 - Present
Led the development of a GPT-4-powered financial assistant using LangChain, integrating machine learning algorithms for object detection and feature extraction, automating over 500 monthly research reports, and accelerating decision cycles. Built real-time fraud detection APIs with FastAPI deployed on Kubernetes clusters achieving sub-second scoring and processing over 1 million daily transactions. Designed scalable ML pipelines with GCP Dataflow and BigQuery for geospatial data processing, reducing latency by 50%. Automated regulatory report drafting using GPT-4 and LangChain, cutting compliance preparation time by 60% while improving audit traceability and accuracy. Established robust CI/CD pipelines with GitHub Actions, Docker, and Kubernetes, reducing deployment cycles from 2 weeks to 3 days and improving rollback safety and release stability for mission-critical applications.
Senior AI/ML Engineer at Charter Communications
May 31, 2024 - August 23, 2025
Developed and deployed a Gen AI virtual customer assistant using LangChain and GCP Vertex AI that handled over 10,000 daily queries with 99.98% uptime. Engineered advanced NLP log analytics pipelines using Python, spaCy, and FastAPI integrating geospatial data mapping to reduce technician truck rolls by 22% via predictive maintenance alerts. Built scalable microservices backend with Docker and GKE, enabling seamless omnichannel Gen AI access across platforms with 55% reduced latency. Designed synthetic data generators using GANs to augment chatbot datasets, improving model robustness and reducing error rates by 30%. Integrated FAISS and Pinecone vector similarity search over 10M+ knowledge base documents, boosting Gen AI recommendation precision to 95% for enhanced user experience and engagement.
Data Engineer at Fortis Healthcare
December 31, 2022 - August 23, 2025
Designed and optimized large-scale ETL pipelines for ingesting, cleaning, and indexing clinical literature and geospatial data sets, reducing query latency by 40% and accelerating research workflows. Built automated data workflows consolidating patient scheduling datasets producing accurate deliverables for no-show detection supporting $200K+ in annual savings. Engineered document ingestion pipelines using OCR (Tesseract) and OpenCV applying object detection for healthcare records, reducing claims processing time from 5 days to 2 hours. Deployed containerized data processing services on Docker, Kubernetes, and Vertex AI Pipelines, integrating point cloud workflows to improve scalability and reduce deployment cycles by 70%.
Data Engineer at Lincoln Financial Group
May 31, 2020 - August 23, 2025
Developed enterprise-graded data governance workflows for bias detection and compliance audits ensuring accurate deliverables across credit and loan datasets. Built real-time ingestion and transformation pipelines using Apache Kafka and Spark for transactional data enabling overdraft risk analytics and reducing potential financial losses by $1.5M annually. Architected streaming pipeline integrating computer vision preprocessing (OpenCV) with laser scanning inputs enhancing anomaly detection accuracy and lowering false positives. Automated customer complaint classification with NLP pipelines (spaCy, NLTK) applying feature extraction to process 50,000+ tickets monthly reducing review time by 50%.
Data Analyst at Mutex Soft Solutions
January 31, 2018 - August 23, 2025
Developed and documented comprehensive data analysis workflows using Jupyter Notebooks incorporating mapping of axis-aligned datasets for improved collaboration and knowledge sharing. Extracted, transformed, and optimized relational datasets using SQL and PL/SQL ensuring accurate deliverables supporting data-driven decision-making across departments. Utilized Python libraries such as Pandas and NumPy for preprocessing and feature extraction significantly improving data accuracy, integrity, and operational efficiency. Conducted in-depth exploratory data analysis (EDA) to identify critical patterns, trends, and anomalies delivering actionable insights that supported strategic initiatives and business growth opportunities. Implemented robust data quality assurance and validation protocols including optimized SQL queries and stored procedures ensuring high consistency and reliability of enterprise datasets for reporting and compliance purposes.

Education

Master's at University of North Texas
January 11, 2030 - August 23, 2025

Qualifications

Add your qualifications or awards here.

Industry Experience

Financial Services, Healthcare, Software & Internet, Professional Services, Government