I am a data engineer with 5+ years of experience designing and implementing big data and cloud-native data platforms. I specialize in building scalable data pipelines, managing Hadoop/Spark ecosystems, and delivering analytics-ready data for business intelligence and data science use cases. I thrive in collaborative, Agile environments and enjoy turning complex data problems into reliable, repeatable solutions. In my roles across banking, tech and enterprise contexts, I have designed end-to-end ETL/ELT pipelines, migrated on-premises data to cloud data lakes and warehouses, and implemented real-time streaming architectures. I am motivated by optimizing performance, ensuring data quality, and delivering actionable insights through dashboards and reports.

Jashwanth Reddy

I am a data engineer with 5+ years of experience designing and implementing big data and cloud-native data platforms. I specialize in building scalable data pipelines, managing Hadoop/Spark ecosystems, and delivering analytics-ready data for business intelligence and data science use cases. I thrive in collaborative, Agile environments and enjoy turning complex data problems into reliable, repeatable solutions. In my roles across banking, tech and enterprise contexts, I have designed end-to-end ETL/ELT pipelines, migrated on-premises data to cloud data lakes and warehouses, and implemented real-time streaming architectures. I am motivated by optimizing performance, ensuring data quality, and delivering actionable insights through dashboards and reports.

Available to hire

I am a data engineer with 5+ years of experience designing and implementing big data and cloud-native data platforms. I specialize in building scalable data pipelines, managing Hadoop/Spark ecosystems, and delivering analytics-ready data for business intelligence and data science use cases. I thrive in collaborative, Agile environments and enjoy turning complex data problems into reliable, repeatable solutions.

In my roles across banking, tech and enterprise contexts, I have designed end-to-end ETL/ELT pipelines, migrated on-premises data to cloud data lakes and warehouses, and implemented real-time streaming architectures. I am motivated by optimizing performance, ensuring data quality, and delivering actionable insights through dashboards and reports.

See more

Experience Level

Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Expert
Intermediate
Intermediate
See more

Language

English
Fluent

Work Experience

Data Engineer at Scotia Bank
May 1, 2024 - November 21, 2025
Designed and implemented ETL pipelines in AWS Glue to ingest and transform data from external sources (S3, ORC/Parquet/Text) into Amazon Redshift. Implemented real-time data processing using Kafka and Spark Streaming, consuming from Kafka topics and pushing results into HBase and Oracle databases. Built and maintained Spark applications using Spark SQL on Databricks; automated data workflows using Apache Airflow and Oozie. Developed data ingestion using Kafka and integrated various data sources into Hadoop and Cassandra. Used Sqoop for transferring data between HDFS and relational databases; designed Hive queries with partitioning and bucketing to optimize performance. Leveraged Talend for loading data into Hive; engineered Snowflake pipelines, and automated data movement to S3/Redshift. Deployed CloudFormation templates, managed Kafka/Zookeeper clusters, and implemented security practices (MTLS, OAuth, IP allowlists).
Senior Data Engineer at Acko
August 1, 2023 - August 1, 2023
Built real-time data pipelines using Apache Kafka and Spark Streaming for high-velocity data. Implemented Azure-based ingestion and transformation via Azure Data Factory and Azure Databricks; orchestrated end-to-end ETL/ELT pipelines with scheduling. Migrated on-prem SQL Server/Oracle to Azure Synapse Analytics and Azure SQL Database; moved MongoDB to Oracle for DSS-style analytics. Utilized SSIS for robust ETL from multiple sources to targets. Migrated datasets from Oracle/SAS to Hive and Azure Data Lake; implemented Kafka producers/consumers with Spring Kafka. Created data services ecosystems across relational, NoSQL, and big data technologies; automated workflows with Airflow and Crontab. Developed ML models in Python/PySpark for classification; loaded data into Snowflake and visualized in Power BI; maintained documentation in Confluence and tracked with JIRA. Employed Kubernetes and Docker for CI/CD, practiced BDD/TDD, and automated infrastructure with PowerShell and ARM templates;
Data Engineer/Hadoop Developer at Adobe
February 1, 2021 - February 1, 2021
Designed end-to-end ETL pipelines; migrated on-premises data to Amazon Redshift; built Kafka consumer APIs in Scala for streaming and batch processing in HDFS, Hive, and Impala. Worked extensively with AWS Big Data services (EC2, S3, EMR, Redshift, DynamoDB); installed and configured Hadoop (HDFS/MapReduce) and developed MapReduce jobs in Java and NiFi. Created and deployed Pig UDFs and loaders; developed SSRS reports and SSIS packages; managed Oozie workflows and optimized Hadoop/Hive performance. Built Spark SQL jobs and Spark-based data transformations for analytics; created Tableau dashboards; tuned Redshift queries for faster analytics; managed user access on Tableau Server. Implemented disaster recovery planning and monitoring with CloudWatch/Elasticsearch; designed backup and failover strategies; collaborated with security and operations teams.
Senior Data Engineer at Acko
March 1, 2021 - August 31, 2023
Built real-time data pipelines using Apache Kafka and Spark Streaming for processing high-velocity streaming data. Designed and implemented data pipelines in Azure using Azure Data Factory (ADF) and Azure Databricks for scalable data ingestion and transformation. Developed and deployed PySpark and Spark SQL transformations in Azure Databricks to manage complex business rules and data transformations. Delivered end-to-end ETL/ELT pipelines orchestrated through ADF, with scheduling via Azure Automation Accounts and Tidal Scheduler. Migrated enterprise data from on-prem SQL Server and Oracle to Azure Synapse Analytics (DW) and Azure SQL Database, ensuring optimized query performance and cost-effective storage. Converted data systems from MongoDB to Oracle to improve DSS-style analytical queries. Utilized SSIS for robust data extraction, transformation, and loading from multiple sources to target systems. Migrated and integrated datasets from Oracle and SAS to Hive and Azure Data Lake for
Data Engineer/Hadoop Developer at Adobe
June 1, 2018 - February 28, 2021
Designed and implemented end-to-end ETL solutions and automated operational workflows using SSIS, AWS Data Pipeline, and NiFi, enabling robust and scalable data processing. Migrated on-premises databases and structured datasets to Amazon Redshift, implementing JSON schema mappings for structured ingestion from S3. Built data ingestion pipelines using Kafka consumer APIs in Scala, with downstream processing in HDFS, Hive, and Impala. Worked extensively on Big Data processing on AWS using EC2, S3, EMR, Redshift, and DynamoDB to store and process large volumes of structured and semi-structured data. Installed and configured Hadoop (HDFS, MapReduce) and developed multiple MapReduce jobs in Java and NiFi for data pre-processing. Developed and deployed custom Pig UDFs and loaders to manipulate and transform data as per business requirements. Designed SSRS reports and developed SSIS packages for seamless data movement and transformation from heterogeneous sources like Teradata, Oracle, and SQ

Education

Post Graduation Diploma in Project Management at Fleming College
January 11, 2030 - April 1, 2024
Bachelor of Engineering at JNTUH
January 11, 2030 - May 1, 2018
Post Graduation Diploma in Project Management at Fleming College
January 11, 2030 - April 1, 2024
Bachelors in Engineering at JNTUH
January 11, 2030 - May 1, 2018

Qualifications

Add your qualifications or awards here.

Industry Experience

Software & Internet, Computers & Electronics, Professional Services, Financial Services