Available to hire
I am a results-oriented Data Engineer with over four years of experience building and optimizing large-scale data pipelines and analytics solutions. I specialize in Azure data services (ADF, Synapse, Data Lake, SQL Database) and Databricks to design scalable ETL/ELT workflows.
I excel in data modeling, warehousing, and big data processing with PySpark, SQL, and Python. I enjoy integrating diverse data sources, maintaining data quality, and delivering cloud-based solutions that empower business users and drive informed decision-making. I’m a collaborative team player who partners with analysts and stakeholders to transform data into actionable insights.
Skills
Work Experience
Data Engineer at Brooks Running
May 1, 2024 - November 3, 2025Designed and maintained end-to-end data pipelines using Azure Data Factory and Databricks to process large volumes of structured and semi-structured data. Developed scalable ETL/ELT workflows integrating data from multiple sources into Azure Data Lake and Synapse Analytics for downstream analytics. Built and optimized PySpark scripts for data transformation, cleansing, and enrichment in Databricks notebooks. Implemented data models using star and snowflake schemas in Synapse and Snowflake to support reporting and self-service BI. Created and managed Azure SQL Databases to store and serve high-quality curated datasets for internal applications and reports. Ensured data quality and consistency by applying validation rules, monitoring pipeline health with Azure Monitor, and handling exceptions with alerts and logging. Collaborated with business analysts and stakeholders to gather requirements and deliver insights via Power BI dashboards and curated datasets. Used Git and Azure DevOps for
Data Engineer at NeenOpal Intelligent Solutions
July 1, 2023 - July 1, 2023Spearheaded the development and implementation of Snowflake-based data pipelines, automating processes and reducing manual intervention by 99%, while empowering business users with enhanced data ownership. Deployed Streamlit applications in Python on AWS EC2, resulting in a 90% reduction in time expended for business users. Implemented DELSERT logic in SQL to dynamically handle primary key/composite key updates, reducing execution time by 50% and enhancing stability for data deletion and insertion operations. Revamped billing automation pipelines leveraging APIs and Selenium for major retailers like Amazon, Google, Apple, and Kobo using Airflow, achieving a 40% surge in efficiency and a 60% reduction in labor intensity. Orchestrated a data scraping mechanism via Scrapy to acquire critical data of over 100,000 books from Amazon, ensuring precise campaign discount information and enhanced retailer data validation. Snowflake, DBT, Airflow, Python, SQL, Scrapy. Led the development of an An
Data Engineer at Express Scripts, India
March 1, 2022 - March 1, 2022Assisted in designing, building, and maintaining scalable data pipelines using Azure Data Factory and Databricks. Ingested, cleaned, and transformed structured and unstructured data from multiple sources into centralized data platforms. Supported the development of ETL/ELT workflows to ensure reliable and efficient data processing. Worked with Azure Data Lake, Azure Synapse, and SQL databases to store and manage large datasets. Implemented basic data validation, quality checks, and monitoring to ensure accuracy and consistency of data. Collaborated with data scientists, analysts, and business teams to make data available for reporting and analytics. Learned and applied best practices in cloud security, performance tuning, and cost optimization for Azure resources. Assisted in creating documentation for data pipelines, workflows, and architecture.
Education
Master of Science (M.S) in Information Technology at Kennesaw State University
May 1, 2023 - May 1, 2025Qualifications
Industry Experience
Software & Internet, Retail, Professional Services
Skills
Hire a Data Scientist
We have the best data scientist experts on Twine. Hire a data scientist in Smyrna today.