Shubham Gagrani is a Data Engineer with 5. 5+ years of hands-on experience in building and maintaining data pipelines using Python, Apache Spark, and AWS services.
Designed and maintained diverse data pipelines handling large amounts of data.
Increased pipeline speeds by up to 10x and reduced errors by 20 percent.
Successfully set up and managed AWS services, implementing CI/CD pipelines.
Increased pipeline speeds by up to 10x.
Reduced errors by 20 percent in data extraction and storage processes.
Attracted 235 registered users to a SaaS web application.
Overview: Fine-tuned a Stable Diffusion model to generate artwork in the traditional Japanese Ukiyo-e style. Responsibilities: Developed an AI Model by fine-tuning a Stable Diffusion model on Japanese Ukiyo-e style images.
Key outcomes:
Optimized training using 33 instance images and 1,000 regularization images.
Overview: A web application to generate sound samples from text prompts, deployed on Render. Responsibilities: Developed and deployed a SaaS application using Flask, Bootstrap, MySQL, and replicate.ai for AI model inference.
Key outcomes:
Successfully attracted and managed 235 registered users.
Ensured reliable performance and security through secure payment processing.
Overview: Zid is a cutting-edge e-commerce platform designed for SaaS applications. Responsibilities: Developed data pipelines using the medallion architecture for analytics. Maintained MySQL databases, automated tasks, and monitored AWS Glue jobs.
Key outcomes:
Ensured smooth operation of MySQL databases through maintenance and performance optimization.
Implemented automation strategies to streamline repetitive tasks.
Overview: Velocity Media Data Aggregator is an innovative project for extracting, integrating, and presenting public information about places in South Africa. Responsibilities: Developed and maintained data extraction scripts using Apify and stored data in AWS S3.
Key outcomes:
Developed and executed comprehensive testing plans to ensure data accuracy.
Successfully set up and managed AWS services, implementing CI/CD pipelines.
Overview: KI Immo is a real estate ads aggregator website that collects and normalizes ads from diverse sources. Responsibilities: Extracted data from 15+ diverse sources using SQL connections, Selenium, Airflow, Spark jobs, Kafka, and APIs.
Key outcomes:
Increased pipeline speeds by up to 10x by transitioning to a different data extraction framework.
Reduced errors by 20 percent by applying a data quality check mechanism.
Shubham Gagrani
Python Developer