Deepali is a Data Engineer with 5. 5+ years of proven experience in optimizing data pipelines and managing cloud services.
Proven ownership in migrating and optimizing data pipelines to AWS Step Functions.
Demonstrated expertise in Redshift cluster management, including encryption and WLM settings.
Consistent recognition for contributions and customer obsession with multiple awards.
Managed encryption across five critical Redshift clusters, minimizing business disruptions.
Successfully migrated all pipelines to Step Functions seamlessly without errors, ensuring uninterrupted functionality.
Spearheaded the migration of S3 files to Glacier for enhanced cost efficiency and S3 management.
Overview: Currently focused on creating and optimizing data pipelines for efficient data processing. Responsibilities: Created and optimized data pipelines, contributing to efficient data processing. Developed and tuned SQL queries to enhance performance in data processing. Managed and maintained Redshift databases to ensure optimal performance and reliability.
Key outcomes:
Created and optimized data pipelines for efficient data processing.
Developed and tuned SQL queries for performance enhancement.
Managed and maintained Redshift databases to ensure optimal performance.
Overview: Managed encryption across five critical Redshift clusters and orchestrated synchronization for production clusters. Responsibilities: Managed encryption across five critical Redshift clusters, collaborating to minimize business disruptions. Orchestrated synchronization for two production clusters, ensuring seamless encryption and smooth transition of WBR jobs. Proposed and implemented migration of all pipelines to Step Functions seamlessly, ensuring uninterrupted functionality.
Key outcomes:
Managed encryption across five critical Redshift clusters minimizing business disruptions.
Successfully migrated all pipelines to Step Functions seamlessly without errors, ensuring uninterrupted functionality.
Overview: Responsible for data extraction, transformation, and production job management within a Big Data environment. Responsibilities: Extracted and Imported data from Salesforce to DataLake using Teradata Hadoop Connector, ensuring seamless integration. Transformed data according to CDC logic, maintaining historical data using DataFrame in PySpark. Automated the Salesforce Migration component using UNIX shell scripting for efficiency.
Key outcomes:
Automated Salesforce Migration component using UNIX shell scripting for efficiency.
Optimized HiveQL queries for ORC transactional tables, enhancing query performance and efficiency.
Overview: Implemented data ingestion pipelines from various sources into a DataLake using CDC logic. Responsibilities: Implemented data ingestion pipelines to extract data from various sources (SFDC, SAP, VISTAAR, EMPOWER, BRAZIL) and loaded it into the A3 DataLake using CDC logic. Utilized Sqoop for importing and exporting data between relational database systems (SQL Server, Oracle) and HDFS.
Key outcomes:
Implemented data ingestion pipelines from various sources using CDC logic.
Optimized Spark code performance by implementing best practices.
Deepali
Big Data Engineer