Sakshi is a Data Engineer with 5. 9 years of experience specializing in Azure Data Engineering.
Designed and built ETL pipelines for data ingestion and transformation
Achieved a 30% reduction in processing time through optimization strategies
Increased data throughput by 25% with performance tuning
Implemented CI/CD pipelines using Jenkins for automated deployment
Developed robust data validation frameworks to ensure high data quality
Achieved a 30% reduction in processing time by optimizing data transformations
Increased data throughput by 25% through pipeline performance tuning
Reduced redundant information by 20% by creating and optimizing SQL-based transformations
Overview: Developed a data integration platform to consolidate and transform data from multiple sources. Responsibilities: Designed and built ETL pipelines in Azure Data Factory for data ingestion and transformation. Used Databricks and PySpark to handle large datasets, applying data cleansing and transformation logic. Implemented SQL stored procedures for processing and storing data. Set up CI/CD pipelines in Jenkins for automated deployment and monitoring of data workflows.
Key outcomes:
Designed and built ETL pipelines for data ingestion and transformation
Implemented SQL stored procedures for data processing and storage
Set up CI/CD pipelines for automated deployment and monitoring
Overview: Developed a predictive data platform for ingesting, processing, and analyzing operational data. Responsibilities: Developed ETL workflows in Azure Data Factory for ingesting data into Azure SQL and Data Lake. Leveraged PySpark and Databricks for data processing and complex transformations. Integrated Power BI for visualization, enabling real-time data access for business stakeholders.
Key outcomes:
Developed ETL workflows for data ingestion
Leveraged PySpark and Databricks for complex data transformations
Integrated Power BI for real-time data access
Overview: Developed a robust data pipeline to integrate, transform, and analyze large volumes of customer data. Responsibilities: Designed and implemented end-to-end data pipelines using Azure Data Factory and Databricks. Utilized PySpark to optimize data transformations and aggregations, achieving a 30% reduction in processing time.
Key outcomes:
Utilized PySpark to optimize data transformations and aggregations, achieving a 30% reduction in processing time
Developed CI/CD workflows in Jenkins for automated deployment and testing
Overview: Developed a scalable data aggregation platform for centralizing sales data from regional databases. Responsibilities: Built ETL pipelines using AWS Glue and PySpark to extract and transform data from multiple regional databases. Integrated SQL and Redshift for efficient data warehousing and querying.
Key outcomes:
Built ETL pipelines using AWS Glue and PySpark
Created and optimized SQL-based transformations, reducing redundant information by 20%
Sakshi
Azure Data Engineer