Available to hire
I am Priyatham Reddy Mara, a Sr. Data Engineer with over 10 years of IT experience specializing in Big Data Technologies, Hadoop ecosystem, Data Warehousing, and SQL-related technologies. I have hands-on experience with Spark, Python/Scala, and cloud platforms including GCP, Azure, and AWS to deliver scalable data analytics solutions.
I have led data integration and analytics initiatives across healthcare, retail, and financial services, leveraging Informatica, Snowflake, Databricks, Palantir Foundry, and modern lakehouse architectures to enable data-driven decision making.
Language
English
Fluent
Work Experience
GCP Data Engineer at Cardinal Health
April 1, 2024 - November 6, 2025Developed real-time streaming and batch data pipelines using PySpark, Apache Flink, Kafka, and Hive on a distributed Hadoop cluster. Leveraged Spark with Python for analytics and ML workflows, including Vertex AI for scalable inference. Built Python scripts for data ingestion, transformation, and validation; implemented CI/CD pipelines with Git. Tuned Snowflake performance via warehouse sizing, clustering, and query profiling; integrated Oracle HCM Cloud data with BigQuery and Snowflake using SSIS/Informatica and Python APIs. Implemented multi-cloud ingestion pipelines across GCP and AWS, combining PySpark, AWS Glue, and Airflow to support near real-time reporting. Designed RESTful APIs and microservices (Java Spring Boot) to orchestrate ingestion and data sharing. Created Matillion ETL pipelines on AWS/GCP and integrated with AWS Glue and GCP Cloud Composer to automate complex workflows. Built dashboards in Foundry Workshop and Power BI for executive oversight; delivered AI-enabled an
Sr. Data Engineer at The Home Depot
March 1, 2024 - March 1, 2024Designed and implemented robust GCP-based data solutions using Compute Engine, Cloud Storage, SQL Server, load balancers, and auto-scaling. Wrote Spark SQL data ingestion scripts from Oracle to Spark clusters with complex joins; built distributed systems with Spark and Scala. Developed Ab Initio ETL workflows, and Scala applications for streaming data to MongoDB and HDFS. Built scalable ETL pipelines using Azure Data Factory (ADF), Azure Synapse Analytics, and Data Lake Storage; automated data movement across ADF, SSIS, and Databricks to improve latency. Created REST APIs using Java Spring Boot; integrated Snowflake with Synapse and Databricks for cross-platform analytics; deployed Databricks/Foundry pipelines for retail analytics; implemented CI/CD with Terraform.
Data Engineer at Arvest Bank
November 1, 2023 - November 1, 2023Designed and deployed a robust reporting data warehouse, migrating ETL pipelines from Talend to Informatica; automated ETL using AWS Glue, Lambda, and PySpark. Built Azure Data Factory and Azure Databricks pipelines to ingest structured and unstructured data from multiple sources. Engineered large-scale data processing with AWS EMR, Spark Streaming, and Kafka, delivering data to Redshift, Athena, and S3. Implemented secure data processing on AWS (S3, Lambda, DynamoDB, Glue) and Azure (Blob Storage, Synapse, Data Lake) to meet performance and regulatory needs. Built lakehouse models in Fabric Synapse and integrated Fabric Power BI with OneLake. Monitored pipelines with Azure Monitor/Log Analytics; authored complex SQL; integrated SAP data with Databricks Lakehouse. Implemented Snowflake staging and Bronze/Silver/Gold transformations; deployed MLOps using Jenkins and Terraform; integrated Kafka for real-time streaming.
Data Engineer at Cipla
November 1, 2020 - November 1, 2020Developed Python scripts to automate product catalog management; designed RESTful APIs to integrate catalog with inventory and order systems. Used Pandas and NumPy for data cleaning, transformation, and processing; built data pipelines handling large-scale product and inventory data. Automated reorder point calculations to optimize stock levels. Processed HL7 messages and Epic EHR extracts to support pharma analytics. Rewrote data wrangling scripts in PySpark on Databricks; integrated Python with MySQL and PostgreSQL for catalog data management. Built ETL workflows using SSIS to feed data into AWS Redshift; developed scheduled tasks with Python and Cron, and orchestrated Hadoop jobs with Oozie. Created Snowflake data sharing and time-travel capabilities; built Tableau dashboards for sales and inventory; integrated with AWS Glue/S3/Redshift and Databricks for end-to-end analytics.
Education
Qualifications
Industry Experience
Healthcare, Retail, Financial Services, Life Sciences, Professional Services
Hire a Data Scientist
We have the best data scientist experts on Twine. Hire a data scientist in Dublin today.