SUBHADIP DAS

DevOps Engineer

Bengaluru, Karnataka, India8 yrs 1 mo experience
AI EnabledAI ML Practitioner

Key Highlights

  • Expert in building scalable Big Data solutions.
  • Proven track record in data pipeline development.
  • Strong leadership experience in data engineering teams.
Stackforce AI infers this person is a Big Data Engineer with expertise in retail and healthcare analytics.

Contact

Skills

Core Skills

Data EngineeringBig Data SolutionsTeam LeadershipData MigrationData AnalysisHealthcare Analytics

Other Skills

AWS ServicesAirflowAmazon Elastic MapReduce (EMR)Amazon S3AthenaDBTGCPHadoopHiveLinuxMapReducePredictive AnalyticsPredictive ModelingPythonR

About

In the era where " DATA IS NEW OIL" , Somewhere amid everything I fell in love with DATA and big data ecosystem. A competitive, determined, meticulous and charismatic professional with a dash of inquisitiveness who always try to work beyond obvious. An eloquent, possessing strong problem-solving skills and a quick learner. Having extensive knowledge on design, development and, deployment of Big Data solutions in various ecosystems by leveraging top notch big data tech stacks. Big Data project experience in the following domains. 1) Retail 2) Pharma 3) Real Estate 4) Aviation Tools: 1. Airflow 2. Informatica 3. Talend 4. GIT 5. JIRA 6. Snowflake Technologies & Languages: 1. Python 2. Scala 3. Spark 4. pySpark 5. SQL 6. Hive 7. AWS & Google Cloud

Experience

Google

Senior Cloud Data Engineer

Apr 2025Present · 11 mos · Bangalore Urban, Karnataka, India · Hybrid

Walmart global tech india

Senior Data Engineer

Apr 2024Apr 2025 · 1 yr · Greater Bengaluru Area · Hybrid

  • 1. I am working as a Senior Data Engineer for
  • Walmart’s U.S. Omni Data & Customer Analytics -
  • Marketplace team where we directly interact
  • with business stakeholder to provide them
  • quality DATA for business insights by building
  • lighteing fast, scalable and robust data pipeline
  • 2. I have built a data pipeline for Walmart’s
  • Fulfillment Seller Center team which will help
  • them to track order return related details.
  • Pipeline handles approx 2.5 billion of rows on
  • daily basis which has a inbuilt data quality
  • framework
  • 3. We have used Spark, Scala,Snowflake,GCP,Hive
  • and DBT to build the entire data pipeline
  • 4. Currently product is up and running in
  • production with 99.95% success rate
SparkScalaSnowflakeGCPHiveDBT+2

Tesco bengaluru

Senior Software Development Engineer 2

Mar 2023Dec 2023 · 9 mos · Bengaluru, Karnataka, India · Hybrid

  • 1.Developing an automated data pipeline for our Price , Promotion & Product Business units
  • 2.This pipeline will help business to understand the currents prices for each product we are going to sell under Tesco Stores.
  • 3.Promotion & Product pipeline will help your data science team to understand the cost of products after giving certain promotions or campaigns which will eventually help business to take business decisions at the right time
  • 4.Technologies we are using to build the entire pipeline are: Spark,Hive,Python,AWS Services along with Airflow to automate the process
  • 5.My roles & responsibilities are to develop certain modules along with testing & deployment of the same module to Production environment for end users.
SparkHivePythonAWS ServicesAirflowData Engineering+1

Target

Senior Data Engineer

Mar 2022Mar 2023 · 1 yr · Bengaluru, Karnataka, India

  • I have been working across multiple projects to build pipelines for our Items,Sales, Vendors as well as for carriers data platforms. We are incorporating Monitors and data quality checks in every flows to ensure providing quality data to each end users. Tools we are using to build these flows are Spark, Scala, Hive, and Airflow
SparkScalaHiveAirflowData EngineeringBig Data Solutions

Nike

Data Engineer

Apr 2021Mar 2022 · 11 mos · Bangalore Urban, Karnataka, India

  • Developed an innovative, robust, scalable, and efficient automated and data pipeline using Airflow DAG. Fetched data from the S3 storage layer to implement business logic's on the data set as per business need using SPARK SQL. Implemented S3 Key Sensor logic to sense
  • new file existence in S3 area which will reduce infrastructure cost and loaded the final output data set to different databases (Snowflake & Hive) for the end-users and enabled a Data Quality checks mechanism to the pipeline to understand data quality from the report
AirflowSPARK SQLSnowflakeHiveData EngineeringBig Data Solutions

Mu sigma inc.

2 roles

Big Data Developer Lead

Promoted

Mar 2020Apr 2021 · 1 yr 1 mo

  • 1. Leading a team of 6 junior developers, responsibilities includes design,
  • development and, deployment of Big Data solutions in various
  • ecosystems
  • 2. Developing an end to end data pipeline for Predictive expected
  • delivery date model. Creating Analytical data set by using different
  • types of data set which are coming from different data sources( e.g.
  • Snowflake, Athena, S3) and feeding that ADS to PEDD model and
  • automating the whole pipeline using Airflow DAG
  • 3. Worked collaboratively with a team member in a data migration
  • project where We migrated all business logic's from the Alteryx tool to
  • Spark SQL and automated it using Airflow DAG and built a Tableau
  • dashboard on top of the final data set
  • 4. Optimized pySpark code modules to enhance code performance
  • which helped us to reduce the run time of the automated pipeline
  • which led to the reduction of infrastructure cost of the business unit
  • and employed automated Data Quality report generation mechanism
  • in the pipeline
AirflowSparkSnowflakeAthenaS3Big Data Solutions+1

Big Data Developer

Apr 2018Feb 2020 · 1 yr 10 mos

  • 1. Worked in a data migration project from SAP HANA to Snowflake DB
  • using Talend Big data platform tool. Designed Entity relationship
  • models for Snowflake DB by analyzing SAP HANA flow graphs.
  • Developed a dynamic, scalable, and automated data migration
  • pipeline including error handling mechanism in the migration
  • framework
  • 3. Contributed to a full-fledged project for a fortune 500 conglomerate
  • which includes the ingestion of data from various
  • sources/types/networks to a centralized data mart with auditing and
  • alert mechanism. The project catered to a data-driven decisionmaking
  • approach for the client
TalendSnowflakeData MigrationBig Data Solutions

Capgemini

Data Analyst

Apr 2017Oct 2017 · 6 mos · Bangaon Area, India

  • 1.Being a Data Analyst I was part of a healthcare project. As the name implies, medical and healthcare data analysts use data from a number of sources to assist improve healthcare outcomes. They typically concentrate on the business side of medicine, increasing patient care or simplifying operations
  • 2.Tech Stack used- R,SQL,Python, Tableaue
RSQLPythonTableauData AnalysisHealthcare Analytics

Education

Dr. B.C. Roy Engineering College

Bachelor of Technology - BTech — Electronics and Communications Engineering

Jan 2012Jan 2016

Stackforce found 100+ more professionals with Data Engineering & Big Data Solutions

Explore similar profiles based on matching skills and experience