Hi, I’m Bipin Thapa, an AI/ML Engineer with 6+ years of hands-on experience designing and deploying production-grade AI systems, MLOps pipelines, and scalable backend solutions. I’ve built RAG systems on Bedrock + LangChain + Pinecone/FAISS, automated ML releases with MLflow and CI/CD, and led end-to-end model lifecycles under defined SLAs across financial and insurance domains. I love turning ambiguous product ideas into measurable ML deliverables. My work spans OCR-to-structured pipelines, data/ML orchestration with PySpark and Kafka, cloud-based inference at low latency, and lakehouse patterns with Iceberg/Snowflake. I partner closely with SMEs and product teams, write maintainable code and runbooks, and mentor engineers to deliver reliable, cost-conscious AI solutions.

Bipin Thapa

Hi, I’m Bipin Thapa, an AI/ML Engineer with 6+ years of hands-on experience designing and deploying production-grade AI systems, MLOps pipelines, and scalable backend solutions. I’ve built RAG systems on Bedrock + LangChain + Pinecone/FAISS, automated ML releases with MLflow and CI/CD, and led end-to-end model lifecycles under defined SLAs across financial and insurance domains. I love turning ambiguous product ideas into measurable ML deliverables. My work spans OCR-to-structured pipelines, data/ML orchestration with PySpark and Kafka, cloud-based inference at low latency, and lakehouse patterns with Iceberg/Snowflake. I partner closely with SMEs and product teams, write maintainable code and runbooks, and mentor engineers to deliver reliable, cost-conscious AI solutions.

Available to hire

Hi, I’m Bipin Thapa, an AI/ML Engineer with 6+ years of hands-on experience designing and deploying production-grade AI systems, MLOps pipelines, and scalable backend solutions. I’ve built RAG systems on Bedrock + LangChain + Pinecone/FAISS, automated ML releases with MLflow and CI/CD, and led end-to-end model lifecycles under defined SLAs across financial and insurance domains.

I love turning ambiguous product ideas into measurable ML deliverables. My work spans OCR-to-structured pipelines, data/ML orchestration with PySpark and Kafka, cloud-based inference at low latency, and lakehouse patterns with Iceberg/Snowflake. I partner closely with SMEs and product teams, write maintainable code and runbooks, and mentor engineers to deliver reliable, cost-conscious AI solutions.

See more

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
See more

Language

English
Fluent

Work Experience

Senior AI/ML Engineer at Clear Fork Bank
November 1, 2023 - Present
Architected a production RAG system on AWS Bedrock + LangChain with Pinecone/FAISS, delivering grounded answers with guardrails and retrieval evaluators. Deployed FastAPI inference on GKE with HPA, readiness/liveness probes, and resource tuning to maintain availability and predictable p95 latency. Implemented MLflow experiments and Model Registry with stage transitions; automated rollback on regressions detected in pre/post-deploy quality checks. Built OCR-to-JSON pipeline using OpenCV + Tesseract; reduced manual data entry and downstream error rates in financial workflows. Automated retraining through SageMaker Pipelines triggered by drift thresholds and model performance deltas. Added offline RAG evaluation suites with quality gates in CI before releasing new indices/prompts. Standardized Docker multi-stage builds and Helm releases; generated SBOMs and enforced policy checks as part of CI/CD. Designed Iceberg + PySpark lakehouse federated to Snowflake for BI; enabled time-travel anal
Python Developer at Cardinal Health
September 1, 2021 - October 1, 2023
Built Django REST APIs for core banking workflows with RBAC, audit trails, and idempotency for reliable transaction handling. Optimized ORM/SQL and implemented caching to reduce p95 latency and increase throughput during peak hours on AWS. Split monolithic endpoints into microservices deployed to Kubernetes; used blue/green releases for safer iterations. Implemented async tasks and schedulers for reconciliations, notifications, and reporting to offload real-time paths. Built ETL and data migrations to Snowflake/PostgreSQL with validations and monitoring hooks for data quality. Automated CI/CD with Jenkins/Git; integrated unit, integration, and contract tests to prevent regressions. Added CloudWatch/Prometheus observability with structured logs and alerting on SLO breaches. Instrumented API usage analytics, quotas, and client keys for visibility and control across internal consumers. Conducted load testing with realistic datasets and guided DB tuning/caching decisions prior to releases.
Python Developer at BCBS Global Solutions
March 1, 2020 - August 1, 2021
Shipped Django/Flask services and REST/SOAP integrations to unify partner systems under consistent contracts. Containerized legacy apps, standardized Helm charts and environment configs for Kubernetes. Built serverless jobs with Lambda and EventBridge for bursty workloads to reduce baseline cost. Implemented Glue/Step Functions pipelines for ingestion/transform with schema checks and alerting. Established test baselines and coverage gates, reduced hotfixes via pre-merge quality checks. Hardened public endpoints with API Gateway throttles and WAF rules. Centralized secrets and parameterized configs across staging and production. Optimized JSON serialization and pagination to reduce payload size and latency. Added distributed tracing and correlation IDs for faster cross-service debugging. Exported curated datasets to Snowflake/BigQuery for downstream analytics and reporting. Automated lint/type/security scans pre-merge for consistent code quality. Defined SLAs, acceptance tests, and usag
Python Developer at Piezos Technology Solutions Inc
January 1, 2019 - February 1, 2020
Developed a CRUD-based web application using Python, Django, and SQLite with user authentication and role-based access control. Created RESTful APIs for data exchange between frontend and backend services. Automated data extraction and processing tasks using Python scripts and Pandas, reducing manual effort by 70%. Built ETL pipeline to load and transform CSV data into PostgreSQL database for reporting and analytics. Designed data visualization dashboards using Matplotlib and Seaborn for actionable insights. Built Flask-based microservice to consume third-party APIs and return processed data in JSON format. Implemented JWT authentication for securing API endpoints and user sessions. Deployed Django application on AWS EC2 with Nginx and Gunicorn for production readiness. Used Git and GitHub for version control and collaborated using Agile methodology. Wrote unit tests with PyTest and automated API testing with Postman to ensure code quality.

Education

Bachelor degree in computer science at Sangai international college
January 11, 2030 - January 5, 2026

Qualifications

Add your qualifications or awards here.

Industry Experience

Financial Services, Professional Services, Software & Internet, Healthcare, Life Sciences, Other