Mayank Pandharpurkar

CEO

Bengaluru, Karnataka, India11 yrs 6 mos experience
Highly Stable

Key Highlights

  • Led successful cloud migration projects across multiple platforms.
  • Expert in designing scalable data architectures for high-volume environments.
  • Proficient in ETL development and performance optimization.
Stackforce AI infers this person is a Big Data and Cloud Solutions Architect with expertise in Fintech.

Contact

Skills

Core Skills

Cloud PlatformsBig DataEtl DevelopmentData Warehousing

Other Skills

Amazon Web Services (AWS)AWS LambdaPySparkAzure DatabricksQlik SensePentahoETL ProcessesDataStageUnixAutosysAWS Step FunctionsAmazon ECSAmazon S3RedisDatabricks Products

About

Tech Lead with 11+ years of experience in Information Technology, specializing in Data Warehousing, ETL Technologies, Big Data, and Cloud Platforms. Extensive background in successfully leading cloud migration projects across Azure, AWS and GCP with a strong focus on banking and telecom domains. Expertise in designing, implementing and optimizing data solutions for real time streaming and batch based ETL in high-volume environments, ensuring cost-efficiency and performance. Led the end-to-end migration of a Hadoop ecosystem to Azure Cloud, overseeing the design, architecture, and development of the migration strategy. Spearheaded the integration of big data workloads into Azure, optimizing data processing pipelines, ensuring scalability, and managing resources efficiently across cloud services. Key Skills & Highlights: ●  Cloud & Big Data Expertise: Proficient in Azure Databricks, Snowflake, ADLS, Delta lake, Confluent Kafka and cloud-based technologies for building scalable, high-performance data architectures. ● Regulatory Reporting: Hands-on experience with FINRA reports, including CAT (Consolidated Audit Trail), Rule 606, Rule 605 and other regulatory requirements. ● ETL Development: Expertise in ETL design using Pentaho, Ascential DataStage, and IBM Websphere DataStage, with performance tuning across both cloud and on-prem environments. ● Data Engineering & Processing: In-depth knowledge of the Hadoop ecosystem (HDFS, Spark, Impala, Hive, Sqoop), data storage formats (Parquet, ORC, Avro), and tools for big data processing. ● Data Warehousing & Performance Optimization: Solid understanding of OLAP vs. OLTP system design with a focus on data quality, data integrity, and performance optimization. ● Programming: Advanced proficiency in Python, Shell Scripting, and SQL for data manipulation, automation, and system integration. ●  DevOps & Agile: Experience working in Agile environments with strong knowledge of DevOps practices (CI/CD pipelines using Git, Jenkins, Jira). ●  BI Reporting & Dashboards: Basic hands-on experience with QlikSense for report creation and Qlik NPrinting for report publishing across different business stakeholders. Technology Stack: Cloud Platforms: Azure, GCP, Snowflake, Databricks Databases: DB2, Sybase, Greenplum, MySQL, Hive, BigQuery, Snowflake Big Data Ecosystem: Hadoop (HDFS, Sqoop, Impala, Hive, Spark, MR), YARN ETL Tools: Pentaho (6.1, 7.1, 8.3), DataStage (7.5, 8.5, 8.7) BI Tools: QlikSense, Qlik NPrinting DevOps Tools: GIT, Jenkins, Jira Schedulers: Autosys, Control-M

Experience

Jpmorganchase

Vice President of Software Engineering

Dec 2025Present · 3 mos · Bengaluru · Hybrid

  • Lead Data Platform
Amazon Web Services (AWS)AWS LambdaCloud PlatformsBig Data

Morgan stanley

2 roles

Senior Manager

Jan 2022Nov 2025 · 3 yrs 10 mos

PySparkAzure DatabricksBig DataETL Development

Manager

Jun 2020Jan 2022 · 1 yr 7 mos

PySparkAzure DatabricksBig DataETL Development

Hsbc

Big Data Pentaho Developer

Nov 2016Jun 2020 · 3 yrs 7 mos · PUNE

  • Big Data - ETL Developer HSBC Technology, Pune
  • Core Skills & Expertise
  • 1) ETL Processes in Pentaho:
  • ● Expertise in designing, developing, and optimizing ETL workflows using Pentaho Data Integration (PDI).
  • ● Understanding of extracting, transforming, and loading data from various sources into a target system, ensuring data consistency, integrity, and quality.
  • 2) Big Data Technologies (Hadoop, HDFS, Hive):
  • ● Solid experience working with the Hadoop ecosystem, including HDFS (Hadoop Distributed File System) for storing large volumes of data.
  • ● Proficient in querying data stored in HDFS using Hive, which is useful for data warehousing and querying large datasets with SQL-like syntax.
  • 3) Data Modeling:
  • ● Understanding the principles of Data Warehousing (OLAP) vs Transactional Systems (OLTP) and the design differences between them.
  • ● OLAP (Online Analytical Processing) is optimized for querying large datasets and analyzing historical data.
  • ● OLTP (Online Transaction Processing) is designed for managing transactional data in real-time.
  • ● Experience in designing star and snowflake schemas, fact tables, and dimension tables for effective data modeling in an enterprise data warehouse.
  • 4) SQL Optimization and ETL Performance Tuning:
  • ● Involvement in performance tuning of SQL queries to reduce latency and improve query execution.
  • Optimization of ETL processes, ensuring efficient data loading by reducing bottlenecks and streamlining data processing pipelines.
  • 5) Agile Methodology & DevOps:
  • ● Experience working in Agile environments, with an emphasis on iterative development, continuous feedback, and collaboration with cross-functional teams.
  • ● Familiarity with DevOps principles, focusing on automation, CI/CD pipelines, and version control to enhance the software development lifecycle, particularly in data engineering contexts.
Qlik SensePentahoETL DevelopmentBig Data

Ibm india private limited

ETL Datastage Developer

Jul 2014Oct 2016 · 2 yrs 3 mos · Pune

  • ● Involved in developing BI applications at ETL & DB2 Aggregates layer.
  • ● Has contributed in identifying data inputs from source systems, creating source to target mapping, designing and developing jobs based on the business rules spreadsheets (BRS).
  • ● Created Autosys JILs for automation of datastage jobs.
  • ● Developed Unix scripts for automating datastage jobs through TWS(Tivoli Workload Scheduler) & for different applications that is invoked through datastage jobs.
  • ● Coordinated Unit Testing, Integration Testing and User Acceptance Testing (UAT) with client representative for every code change and enhancement.
  • ● Validating the load process of ETL to make sure the target tables are populated according the data mapping provided that satisfies the transformation rules.
  • ● Successfully tuned existing ETL programs to increase processing and loading efficiency.
  • ● Troubleshooting, resolving and escalating data related issues and validating data to improve data quality from source systems.
  • ● Coordinating with downstream and upstream teams to ensure smooth delivery.
DataStageUnixETL DevelopmentData Warehousing

Education

Shri Guru Gobind Singhji Institute of Engineering and Technology, Vishnupuri, Nanded

B.Tech — Computer Science

Jan 2010Jan 2014

Stackforce found 100+ more professionals with Cloud Platforms & Big Data

Explore similar profiles based on matching skills and experience