Divyansh Gupta

Data Engineer

Bengaluru, Karnataka, India6 yrs 11 mos experience

Key Highlights

  • Expert in developing ML solutions for fraud detection.
  • Successfully modernized data systems for major financial clients.
  • Awarded Visa Innovation Award for feature explainability tool.
Stackforce AI infers this person is a Data Engineering expert in Fintech with a focus on machine learning and big data solutions.

Contact

Skills

Core Skills

Data EngineeringPythonMachine LearningBig Data Analytics

Other Skills

AirflowAmazon Web Services (AWS)Apache SparkBootstrapData AnalysisData WarehousingData lakehouseExtract, Transform, Load (ETL)HDFSHadoopHiveJenkinsMongoDBMySQLNoSQL

About

As a Senior Data Engineer (MLOPS) at Visa (PFI), I work on developing and maintaining data pipelines(feature engineering) and ML solutions to score transactions for RTP A2A to identify Risk and prevent Fraud . I have a Bachelor of Technology degree in Computer Science from GLA University and multiple certifications in Snowflake, Python, and SQL. I have over four years of experience in data engineering, working with various technologies such as SQL, Python, Pyspark, HDFS, MySQL, Git, AWS, and Snowflake. I have successfully delivered data modernization projects for American Express and T. Rowe Price, creating efficient database structures, data schemas, and data integration solutions. I have also contributed to enhancing the data quality and reducing cost for American Express by upgrading their BCV system from SAS to Big Data. I enjoy working with data and solving complex problems using innovative and scalable solutions. I have a strong analytical mindset and a passion for learning new tools and frameworks. I collaborate effectively with cross-functional teams and communicate clearly and professionally. I am motivated by the opportunity to create value and impact for the customers and the business through data engineering.

Experience

6 yrs 11 mos
Total Experience
1 yr 7 mos
Average Tenure
4 mos
Current Experience

Fidelity international

Modern Data Engineer III

Dec 2025Present · 4 mos · Bengaluru · Hybrid

Python (Programming Language)SnowflakeData lakehousePySparkSQLData Engineering+1

Visa

Sr. Data Engineer

May 2024Nov 2025 · 1 yr 6 mos · Bengaluru · Hybrid

  • Project Name: PREDICTIVE FRAUD INTELLIGENCE (PFI)
  • Technologies : SQL,Python, Pyspark, HDFS , Airflow
  • Led feature engineering initiatives to build machine learning models for scoring and predicting fraudulent transactions, serving major clients such as Coelsa, New Pay, BaseLine Model, and Pix.
  • Redesigned the feature engineering pipeline, achieving a 90% improvement in feature delivery performance, which significantly accelerated model training, and positively impacted product timelines.
  • Developed an automated framework for feature creation, eliminating the need for individual code writing, thus enhancing efficiency and scalability.
  • Collaborated with data scientists to understand and implement new feature requirements, ensuring alignment with business objectives.
  • Designed and implemented a feature correctness and testing framework, establishing rules for each feature family to maintain data accuracy and integrity.
  • Improved developer experience, code quality, and modular design through the redesign of key modules, promoting better maintainability, and scalability.
  • Attending daily stand-up meetings to update the onsite and offshore teams.
  • I received the Visa Innovation Award for building the Feature Explainability tool to showcase the lifecycle of a feature in the A2A feature store.
  • Collaborating with cross-functional teams to productionize A2A fraud detection models.
PySparkAirflowMachine LearningSQLPythonData Engineering

Paytm payments bank

Senior Software Engineer

Aug 2023May 2024 · 9 mos · Noida, Uttar Pradesh, India · Remote

  • Project Name : DWH
  • Technologies : SQL,Python, Pyspark, HDFS, MYSQL,EMR,S3,Kafka Consumer
  • Responsibilities:
  • Create and manage the data integration system by building Pyspark batch and streaming data pipelines to ingest raw OLTP data from multiple RDBMS,SFTP and Streaming sources into S3/HDFS , where tables can be created to support the creation of data cuts and data marts for business analytics.
  • Configuring custom scheduler to meet the SLA of all the raw data cuts in automated fashion.
  • Debugging and providing BAU support for Pyspark jobs on iteration basis .
AirflowApache SparkBig Data AnalyticsData EngineeringData WarehousingPySpark+2

American express

Software Engineer II - Big Data

Nov 2021Aug 2023 · 1 yr 9 mos · Gurugram, Haryana, India · Hybrid

  • Product : American Express
  • Technologies : SQL Python, Pyspark, HDFS, MYSQL,Git
  • American Express Company (Amex) is an American multinational Financial services corporation specialized in payment cards ( CORP , SMEs as well as Consumer) . Reading SORs of different Entities from HDFS and receiving Files through SFTP from the product and business team , post that massaging the data using Pyspark and creating DAAP ( Facts and Dimensions ) to be utilised by business team for the reporting needs.
  • Responsibilities:
  • Developing Pyspark jobs based on the business rules provided by PO team to provide DAAP.
  • Upgrading the Existing BCV system from SAS to new Big Data (HDFS) that is further used to incentivise Amex sales personal .
  • Created Key KPIs like Charge Volume for (Consumer and SBS and CORP cards) for AsiaPacific region.
  • Designing Data warehouse schema for payments GCS to outsource Data to variousConsumers for their reporting needs.
  • Implementing business rules in the data to get Insights.
  • Developing SQL queries to perform ETL workSlows.
  • Debugging and providing BAU support whenever required in the PI.
  • Orchestrating Data pipeline using AMEx in-house tool Event Engine as well as Java Bean workSlows.
  • Attending Daily Stand-up meets to update the onsite and offshore team.
  • Migrating jobs from magellan to spark for better optimisation .
  • Developed utility to understand the Sill rate and data quality of the SORs and Dimensions .

Teksystems

3 roles

Software Engineer

Jan 2021Nov 2021 · 10 mos

  • Project Name : Data Modernization Duration : 2+ years
  • Client : T. Rowe price
  • Technologies : AWS, Python, Pyspark, Snowflake, Postgres, DB2
  • T. Rowe Price Group, Inc. is an American publicly owned global asset management Firm that offers funds, advisory services, account management, and retirement plans and services for individuals, institutions, and Financial intermediaries.
  • Aws S3 is a source where will get data files from grand central which will pull data from adobe and salesforce database and these data files consist of data related to android, iMac and web sources. Reading and massaging of data is done by Pyspark jobs and transfer the data to Snowflake.
  • Responsibilities:
  • Developing Pyspark jobs for Ingestion Framework.
  • Implementing business rules in the data to get Insights.
  • Developing SQL queries to perform ETL workflows.
  • Used SNS notification mail whenever job got failed.
  • Launching the framework on Kubernetes cluster.
  • Attending Daily Stand-up meets to update the onsite and offshore team.
  • Creating BDDs for code’s functionality testing.

Associate Software Engineer

Promoted

Aug 2019Jan 2021 · 1 yr 5 mos

Data Analysis and Insights Intern

Jan 2019May 2019 · 4 mos · Hyderabad Area, India

  • Great to learn the core concepts of Data warehousing , ETL , Big data ,AWS along with SQL and python.

Education

GLA University

Bachelor of Technology — Computer Science

Jan 2015Jan 2019

Stackforce found 100+ more professionals with Data Engineering & Python

Explore similar profiles based on matching skills and experience