I'm Jemin Shrestha, a Senior Data Engineer with 5+ years of experience designing and operating scalable, reliable, and high-performance data platforms in production. I specialize in SQL, Python, and PySpark to build real-time and batch pipelines, with hands-on work across AWS, Azure, Snowflake, and Databricks. I have a strong background in data modeling, lakehouse/warehouse architectures, and orchestration with Airflow and Dagster. I'm passionate about data quality, governance, and mentoring data teams in modern engineering practices. I enjoy turning complex data into trusted datasets for analytics, and I thrive in cross-functional environments where I can contribute to end-to-end solutions from ingestion to BI-ready marts.

Jemin Shrestha

I'm Jemin Shrestha, a Senior Data Engineer with 5+ years of experience designing and operating scalable, reliable, and high-performance data platforms in production. I specialize in SQL, Python, and PySpark to build real-time and batch pipelines, with hands-on work across AWS, Azure, Snowflake, and Databricks. I have a strong background in data modeling, lakehouse/warehouse architectures, and orchestration with Airflow and Dagster. I'm passionate about data quality, governance, and mentoring data teams in modern engineering practices. I enjoy turning complex data into trusted datasets for analytics, and I thrive in cross-functional environments where I can contribute to end-to-end solutions from ingestion to BI-ready marts.

Available to hire

I’m Jemin Shrestha, a Senior Data Engineer with 5+ years of experience designing and operating scalable, reliable, and high-performance data platforms in production. I specialize in SQL, Python, and PySpark to build real-time and batch pipelines, with hands-on work across AWS, Azure, Snowflake, and Databricks. I have a strong background in data modeling, lakehouse/warehouse architectures, and orchestration with Airflow and Dagster.

I’m passionate about data quality, governance, and mentoring data teams in modern engineering practices. I enjoy turning complex data into trusted datasets for analytics, and I thrive in cross-functional environments where I can contribute to end-to-end solutions from ingestion to BI-ready marts.

See more

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Expert

Language

English
Fluent

Work Experience

Data Engineer at Rogers Communications (Cybersecurity Department)
January 1, 2025 - September 30, 2025
Led migration of security pipelines to AWS using Kafka, Lambda, API Gateway, and Step Functions; built automated Snowflake compliance workflows with Tasks/Streams. Developed Python-based ETL testing (Pytest, pydeequ), anomaly detection, and EventBridge-triggered Lambda validation, reducing SIEM false positives by 28%. Automated infrastructure with Terraform and deployed CI/CD pipelines via GitHub Actions and Concourse for Docker/ECR workloads.
Data Engineer & Analytics Engineer at Tarsi Group
January 1, 2024 - August 31, 2025
Engineered end-to-end ETL/ELT pipelines using PySpark, Airflow, Python, SQL; migrated legacy SSIS/Talend workflows into Snowflake using Snowpipe, Tasks, and Streams. Designed fact/dimension/aggregation models, CDC frameworks, and metadata-driven ingestion for Jira, Workday, OneStream, PRS, STEAM. Developed APIs with FastAPI + API Gateway + containerized Lambda and implemented Databricks Delta (bronze/silver/gold) with CI/CD orchestration.
Senior Data Engineer at Audit Partnership
March 1, 2023 - August 31, 2023
Engineered ingestion frameworks in ADF, Dagster, PySpark for 100+ suppliers (millions of rows per month), improving throughput and scheduling reliability by 30%. Migrated SSIS to cloud pipelines with monitoring, alerts, and error handling; optimized SQL Server procedures for 25–40% faster execution. Delivered curated marts and BI-ready datasets for Power BI, Looker, Tableau with improved audit accuracy.
Data Engineer at Cotiviti
July 1, 2021 - March 31, 2023
Optimized mission-critical SQL and ETL workflows, reducing runtimes by up to 60% across multiple finance/healthcare workloads. Automated validation and cleansing using Python, cutting manual work by 40%. Enhanced stored procedures and batch systems with improved logging, fault tolerance, and tuning.

Education

Postgraduate in Artificial Intelligence at Georgian College
September 1, 2024 - June 30, 2025
Postgraduate in Big Data Analytics at Georgian College
September 1, 2023 - June 30, 2024

Qualifications

Fabric Data Engineer - DP700
January 11, 2030 - January 8, 2026
Fabric Analytics Engineer – DP600
January 11, 2030 - January 8, 2026
AWS Academy – Cloud Foundations
January 11, 2030 - January 8, 2026
AWS Academy – Data Engineering
January 11, 2030 - January 8, 2026
AWS Academy – Machine Learning Foundations
January 11, 2030 - January 8, 2026
HackerRank SQL Advanced
January 11, 2030 - January 8, 2026
Google Data Analytics
January 11, 2030 - January 8, 2026

Industry Experience

Software & Internet, Telecommunications, Professional Services, Media & Entertainment