Shaifali  ·  Lead GCP Data Engineer  ·  10+ yrs

Principal
10+ years experienceremote
Available within 48 hrs

Proof of scale

50% reduction in data retrieval latency
40% reduction in query execution time
30% improvement in ETL pipeline efficiency
40% reduction in query execution time30% improvement in ETL pipeline efficiency50% reduction in data retrieval latency20% reduction in data storage expenses25% reduction in data processing costs

About Shaifali

Shaifali is a Data Engineer with 10+ years of experience in designing and optimizing scalable data pipelines and data warehouses on Google Cloud Platform. She has a proven track record of improving ETL efficiency and ensuring data security across various domains.

10+ years of commercial experience in

Skills(14)

GCPBigQueryApache AirflowTerraformSQLMongoDBDataflowPythonApache BeamOracleLookerPub/SubStackdriverPrometheus

Why hire Shaifali?

Production deploy authorityMentored juniors

Led infrastructure automation using Terraform for efficient resource management.

Automated pipeline deployment with Apache Airflow for real-time systems.

Achieved measurable improvements in ETL efficiency and data retrieval latency.

Ensured compliance with HIPAA/GDPR across multiple domains.

Designed real-time data ingestion pipelines for fraud detection systems.

Reduced BigQuery query execution time by 40%

Improved ETL pipeline efficiency by 30%

Reduced data retrieval latency by 50%

Led cost-optimization efforts, reducing data storage expenses by 20%

Ensured compliance with healthcare data protection laws (HIPAA, GDPR)

Project highlights(11)

Large-Scale Data Warehouse OptimizationLead Data Engineer

Overview: This project involved designing and implementing a petabyte-scale data warehouse on GCP. Responsibilities: Designed and implemented a petabyte-scale data warehouse on GCP. Optimized BigQuery queries, reducing execution time by 40%. Automated ETL workflows using Apache Airflow for seamless data processing. Established best practices for cost optimization and data security. Ensured compliance with data protection regulations via encryption and PII masking. Led infrastructure automation using Terraform for efficient resource management.

GCPBigQueryApache AirflowTerraform

Key outcomes:

  • Reduced BigQuery query execution time by 40%

  • Successfully designed and implemented a petabyte-scale data warehouse

Cloud Data Migration & ETL OptimizationSenior Data Engineer

Overview: This project focused on migrating on-premise data sources to GCP BigQuery and optimizing ETL processes. Responsibilities: Migrated on-premise data sources (CSV, SQL, Oracle, MongoDB) to GCP BigQuery. Developed optimized ETL pipelines with Dataflow and Apache Beam, improving efficiency by 30%. Designed access control mechanisms for secure data governance. Conducted performance tuning, reducing data processing costs by 25%. Integrated Looker for enhanced data visualization and reporting. Ensured minimal downtime during the migration process.

GCPBigQueryDataflowApache BeamSQLOracleMongoDBLooker

Key outcomes:

  • Improved ETL pipeline efficiency by 30%

  • Reduced data processing costs by 25%

  • Ensured minimal downtime during data migration

Predictive Analytics for Financial DataSenior Data Engineer

Overview: Built an end-to-end data pipeline for predictive analytics on financial transactions. Responsibilities: Built an end-to-end data pipeline for predictive analytics on financial transactions. Designed efficient data models in BigQuery, reducing retrieval latency by 50%. Implemented machine learning workflows for fraud detection using Python. Automated ETL processes for real-time financial data processing. Integrated Looker for business intelligence and financial reporting. Ensured compliance with financial data security standards.

BigQueryPythonLooker

Key outcomes:

  • Reduced data retrieval latency by 50%

  • Successfully implemented machine learning workflows for fraud detection

Real-Time Data Processing SystemData Engineer

Overview: Designed and implemented a real-time data ingestion pipeline and event-driven architecture. Responsibilities: Designed and implemented a real-time data ingestion pipeline using Pub/Sub and Dataflow. Developed event-driven architecture for streaming data processing. Optimized BigQuery storage for efficient data querying. Integrated monitoring solutions using Stackdriver and Prometheus. Automated pipeline deployment using Terraform and Apache Airflow. Ensured fault tolerance and high availability for critical data flows.

Pub/SubDataflowBigQueryStackdriverPrometheusTerraformApache Airflow

Key outcomes:

  • Ensured fault tolerance and high availability for critical real-time data flows

Enterprise Data Lake for Retail AnalyticsData Engineer

Overview: Designed a centralized data lake on GCP for retail analytics. Responsibilities: Designed a centralized data lake on GCP for retail analytics. Developed scalable ETL jobs using Dataflow and Apache Beam. Led cost-optimization efforts, reducing data storage expenses by 20%. Integrated customer behavior analytics into business intelligence dashboards. Improved query performance and data retrieval speeds. Ensured data security through encryption and access control mechanisms.

GCPDataflowApache Beam

Key outcomes:

  • Reduced data storage expenses by 20%

  • Improved query performance and data retrieval speeds

Industry experience

Cybersecurity

8 projects
  • Large-Scale Data Warehouse OptimizationLead Data EngineerGCP · BigQuery · Apache Airflow · Terraform
  • Predictive Analytics for Financial DataSenior Data EngineerBigQuery · Python · Looker
  • Enterprise Data Lake for Retail AnalyticsData EngineerGCP · Dataflow · Apache Beam
  • IoT Analytics PlatformSenior Data EngineerBigQuery · Pub/Sub · Dataflow
  • Supply Chain Data HubData Engineer
  • Healthcare Data Aggregation SystemData Engineer
  • Telecom Customer AnalyticsData EngineerBigQuery
  • Real-Time Fraud Detection SystemData EngineerApache Airflow

HealthTech

1 project
  • Healthcare Data Aggregation SystemData Engineer

Legal Tech

4 projects
  • Large-Scale Data Warehouse OptimizationLead Data EngineerGCP · BigQuery · Apache Airflow · Terraform
  • Predictive Analytics for Financial DataSenior Data EngineerBigQuery · Python · Looker
  • IoT Analytics PlatformSenior Data EngineerBigQuery · Pub/Sub · Dataflow
  • Healthcare Data Aggregation SystemData Engineer

Logistics & Supply Chain

3 projects
  • Large-Scale Data Warehouse OptimizationLead Data EngineerGCP · BigQuery · Apache Airflow · Terraform
  • Supply Chain Data HubData Engineer
  • Real-Time Fraud Detection SystemData EngineerApache Airflow

Ready to work with Shaifali?

Schedule an interview and onboard within 48 hours. No long hiring cycles.

At a Glance

Experience10+ years
Work moderemote
Starting from₹1.6 L/mo
Direct hirePossible
Start within48 hours
From₹1.6 L/ month

Single contract. No agency markup confusion.

Typically responds within 4 business hours.

5-day replacement guarantee
48-hour onboarding, single invoice
Direct chat — no recruiter middleman
Seniority signals
Owns production deploysGreenfield architectSystem ownerMentor / leads juniors
VerifiedVetted by Witarist
Technical skills assessed & verified
Background & identity checked
English communication verified
Ready to onboard in 48 hours

Not sure if this is the right fit?

Tell us your requirements and we'll match you with the best candidates.

Shaifali

Data Engineer