Nitesh Madan is a Big Data Developer with 6+ years of experience in designing, implementing, and supporting big data applications. He has a strong proficiency in AWS, Apache Spark, and Python, and has successfully contributed to projects in both financial services and manufacturing domains.
Designed and implemented Big Data applications using Apache Spark and AWS.
Optimized PySpark scripts for data extraction and reporting.
Ensured data quality for critical payment-processing systems.
Contributed significantly to all project stages from requirements to production support.
Successfully designed + implemented Big Data applications using Apache Spark + Hadoop + AWS
Developed + optimised PySpark scripts for data extraction + cleaning + reporting
Implemented Hive + HBase schemas with partitioning + bucketing for performance
Overview: This project supports Mastercard Inc., the second-largest payment-processing corporation worldwide, by providing services for payment transaction processing. Responsibilities: Understood upstream source nature and business cases to ensure data quality and relevance. Monitored multiple batches and executed various scripts across several servers. Performed initial data loads using Initial_Dataload.sql, checked table structures, and filtered data from files. Created Standard Operating Procedures (SOP) for automation, enhancing operational efficiency.
Key outcomes:
Ensured data quality and relevance by understanding upstream source nature and business cases.
Improved operational efficiency by creating SOPs for automation.
Overview: This project supported UniPol, a global manufacturing company specializing in investment casting technology for automotive and aerospace industries. Responsibilities: Understood upstream source data and business requirements to facilitate data processing. Connected to a PostgreSQL instance using workbench for database operations. Performed initial data loads using Initial_Dataload.sql, checking table structures and sample data. Copied all dependencies to EMR and created HBase tables using Phoenix scripts. Developed Spark jobs to process data, supporting data transformation workflows.
Key outcomes:
Processed data for a global manufacturing company to support investment casting technology.
Successfully integrated PostgreSQL, HBase, and Spark for data warehousing and analysis.
Key outcomes:
Ensured quality assurance standards were achieved by compiling and analyzing statistical data.
Successfully implemented Hive and HBase column family schemas with performance techniques like partitioning and bucketing.
Resolved production defects and contributed to ongoing compliance with quality and regulatory requirements.
Nitesh Madan
Big Data Engineer