Available to hire
I’m Mohsin Zahoor, a Senior Data Engineer with 6+ years of experience designing, developing, and optimizing cloud-based data platforms. I’ve led cross-functional teams, streamlined ETL/ELT processes, and enabled real-time analytics across Fintech, Healthcare, Telecom, and Proptech.
I specialize in Python, PySpark, SQL, Snowflake, Redshift, AWS and Azure, Kafka, and a broad set of data technologies. I build scalable data models, automate pipelines, enforce security and governance (GDPR, HIPAA, SOC 2), and collaborate across stakeholders to deliver value.
Skills
See more
Experience Level
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Language
English
Fluent
Work Experience
Lead Data Engineer at Analytiverse
September 1, 2024 - November 21, 2025Led the design and delivery of a multi-layered data platform on AWS (S3, Glue, Lambda, EMR, Redshift) integrated with Snowflake, enabling a shift from batch reporting to real-time analytics. Collaborated with executives and product owners on data strategy, ensuring alignment with business and analytics goals. Managed data ingestion from 15+ sources (APIs, streaming, enterprise databases) using AWS services, ensuring reliable CDC pipelines and schema evolution. Implemented data reliability frameworks with dbt tests, Great Expectations, and CloudWatch, reducing undetected data issues by 40%. Led data governance initiatives, enforcing role-based access, AWS KMS encryption, and automated compliance checks for SOC2 and GDPR. Coached engineers on ELT, CI/CD (GitHub Actions), and Terraform deployments, fostering a culture of ownership and excellence. Partnered with data science teams to operationalize ML pipelines on AWS (SageMaker, Snowflake feature store), reducing model deployment time fro
Data Engineering Team Lead at Cityscape Technology Corporation
September 1, 2024 - September 1, 2024Led a team of 8+ data engineers, providing mentorship, technical guidance, and overseeing successful delivery of complex cloud-native data solutions. Designed and optimized large-scale ETL/ELT pipelines using PySpark, DBT, Delta Lake, Snowflake, and Redshift, improving data quality, consistency, and processing efficiency. Built high-performance platforms handling 1,000,000+ daily API requests on AWS (EC2, S3) and managed cloud infrastructure using Lambda, Airflow, Terraform, and Jenkins. Implemented data governance and security practices to ensure compliance with GDPR, HIPAA, and SOC 2 standards. Spearheaded migration projects and cloud adoption, moving on-premise systems to AWS and enabling scalable, cost-efficient architectures. Collaborated with stakeholders and SCRUM teams to align data strategies with business goals, delivering actionable insights and supporting analytics/ML initiatives. Adopted emerging technologies and best practices, driving innovation, continuous improvement,
Senior Data Engineer at Odyssey Solutions
June 1, 2023 - June 1, 2023Integrated 9+ systems for a fintech client, leveraging Kafka, Debezium, and PySpark Streaming for CDC from SQL Server, feeding batch and real-time analytics pipelines. Built scalable ETL pipelines with PySpark, AWS Glue, Redshift, and Looker, supporting dashboards, reporting, and ML/LLM model training. Implemented Delta Lake on Databricks and MiniO-based data lake solutions, unifying batch and streaming data for transactional consistency. Led architecture and deployment for platforms in commodities trading and crypto intelligence, utilizing AWS services (S3, Lambda, Glue) and Greenplum for AI model support. Developed modular pipelines and contributed to forecasting platforms (OdyxyHat, Coinex), enabling data-driven insights and operational efficiency. Applied modern data engineering techniques to optimize pipeline performance, improve data quality, and streamline processing across multiple domains.
Data Engineer II at Afiniti
July 1, 2022 - July 1, 2022Built and automated robust ETL pipelines using Python, PySpark, Talend, and Snowflake, enabling near real-time ingestion of structured and unstructured data. Led migration of legacy reporting data from MySQL to Greenplum, reducing report generation time from 1 week to 1 day and ensuring data consistency. Developed ingestion pipelines for healthcare patient vitals and device telemetry using AWS (Kinesis, Lambda, Glue), with dynamic transformations and cataloging via Crawlers. Implemented matching logic and automated analytics, improving data accuracy, client reporting, and contributing to revenue growth across APAC and UK clients. Processed HIPAA-compliant healthcare data and implemented SCD transformations for historical accuracy. Built Python sensors and GitHub Actions workflows to automate pipeline deployments, reduce manual intervention, and accelerate release cycles. Coordinated with AI teams and business stakeholders to improve predictive models, streamline CRM processes, and reso
Lead Data Engineer at Analytiverse
July 1, 2023 - PresentLed a team of 8 engineers to architect cloud-native systems and establish organization-wide data governance. Built and scaled a real estate platform handling 1M+ API requests per day on AWS, and optimized ETL pipelines with PySpark, CockroachDB, and Snowflake CDC. Increased data accuracy by adopting Delta Lake; oversaw Airflow, Lambda, and Jenkins with ELK-based monitoring; migrated select on-prem workloads to AWS; consolidated datasets from ITSOs, CREA, TREB into a centralized repository; drove Agile delivery with JIRA; enforced GDPR, HIPAA, and SOC 2 controls. Implemented modular dbt transformations for maintainable data models and supported Databricks PySpark processing.
Senior Data Engineer at Odyssey Solutions
July 1, 2022 - June 20, 2023Integrated nine financial and transactional systems into a real-time data platform using Kafka, Debezium, and PySpark. Coordinated Debezium with Kafka for real-time CDC from SQL Server databases; used PySpark Streaming to process CDC events; built ETL pipelines for financial analytics and dashboards with Looker using PySpark, AWS Glue, and Redshift. Created a hybrid financial data warehouse ingesting data automatically using MinIO and AWS Glue. Configured Spark Streaming for near real-time processing and contributed to no-code forecasting and crypto analytics platforms.
Data Engineer II at Afiniti
June 1, 2019 - July 20, 2022Directed migration of legacy healthcare reporting workloads from MySQL to Greenplum; automated ETL workflows with Talend and PySpark; implemented SCD logic and Python-based sensors for reliable orchestration. Engineered secure Snowflake data pipelines; created near real-time streaming ingestion for patient vitals and device telemetry via Kinesis and Lambda. Built GitHub Actions CI/CD pipelines for Snowflake and AWS ETL deployments; processed HL7/EHR datasets with AWS Glue dynamic frames and PySpark.
Education
Qualifications
Bachelor's Degree in Computer Science
January 11, 2030 - November 21, 2025Bachelors of Computer Science
January 11, 2030 - January 7, 2026Bachelors of Computer Science
January 11, 2030 - January 29, 2026Industry Experience
Financial Services, Healthcare, Telecommunications, Real Estate & Construction, Software & Internet
Skills
See more
Experience Level
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Hire a Data Analyst
We have the best data analyst experts on Twine. Hire a data analyst in Lahore today.