Babitha Sanimaina

DevOps Engineer

Bengaluru, Karnataka, India11 yrs 5 mos experience
Most Likely To Switch

Key Highlights

  • Expert in building scalable data architectures.
  • Proven track record in optimizing data processing workflows.
  • Extensive experience across multi-cloud data ecosystems.
Stackforce AI infers this person is a Data Engineering expert with extensive experience in Retail and E-commerce sectors.

Contact

Skills

Core Skills

Data EngineeringApache Spark

Other Skills

Apache Spark StreamingApache KafkaAmazon Web Services (AWS)Extract, Transform, Load (ETL)Data ArchitectureTableauSpark with ScalaAWS EMRS3ScalaPrestoMicrosoft AzureAzure DatabricksDatabricks SQLSpark

About

Data Engineering leader with 14+ years of experience building scalable data platforms and distributed data processing systems across global organizations. I specialize in designing and operating large-scale data architectures that power analytics, machine learning, and data-driven products. My work has focused on building reliable, high-performance data pipelines and platforms using technologies such as Apache Spark, Databricks, Kafka, and modern lakehouse architectures. Over the years I’ve worked across multiple domains including retail, e-commerce, and travel, contributing to data platform initiatives at organizations such as Tesco, Meesho, Expedia, Walmart, and FIS. My experience includes: • Designing enterprise data platforms using lakehouse and medallion architecture patterns • Building large-scale batch and streaming data pipelines using Spark and Kafka • Developing reusable data ingestion frameworks and platform components • Enabling analytics and machine learning teams with high-quality curated datasets • Improving reliability and scalability of distributed data pipelines • Mentoring engineers and guiding architectural decisions across teams I have extensive experience working with multi-cloud data ecosystems across AWS, Azure, and GCP, designing solutions that balance scalability, performance, and operational reliability. I’m passionate about solving complex data challenges, building robust data platforms, and helping teams unlock the value of data through scalable engineering solutions. Always interested in discussions around data platform architecture, distributed systems, and modern data engineering practices.

Experience

11 yrs 5 mos
Total Experience
1 yr 10 mos
Average Tenure
2 yrs
Current Experience

Tesco bengaluru

Lead data engineer

May 2024Present · 2 yrs · Bengaluru, Karnataka, India · Hybrid

  • Leading workload forecasting pipelines, processing a near real-time data to predict staffing needs at 15-minute intervals across multiple regions and channels of Tesco.
  • This forecasts help reduce costs, improve efficiency, and enhance customer satisfaction by aligning staffing levels with demand
Apache SparkApache Spark StreamingApache KafkaData Engineering

Expedia group

Senior Data Engineer

Apr 2023May 2024 · 1 yr 1 mo · Bengaluru, Karnataka, India

  • Data Engineer | Clickstream Data Enablement Team, Expedia - Present
  • Lead data engineering efforts within the Clickstream Data Enablement Team, focusing on handling large volumes of data to enable seamless user access.
  • Develop and deliver data products tailored for search attribution, contributing to the resolution of complex problems and enhancing overall data accessibility.
  • Drive the design and implementation of robust data pipelines using Spark with Scala to ensure efficient processing and availability of data products.
  • Utilize Tableau for dashboarding, providing visual representations of data insights to support decision-making.
  • Implement EG Data Quality (DQ) checks and validation procedures to ensure the accuracy and reliability of data products.
  • Proficient in AWS EMR and S3, leveraging cloud services for scalable and cost-effective data processing and storage.
  • Collaborate with cross-functional teams to gather and analyze requirements, aligning data solutions with business objectives.
  • Optimize data processing workflows to enhance performance and reduce latency, resulting in a more responsive data infrastructure.
  • Achievements:
  • Successfully led initiatives to enhance data accessibility, providing users with seamless access to data products related to search attribution.
  • Played a key role in solving complex problems through innovative data engineering solutions.
  • Recognized for optimizing data processing workflows, resulting in improved overall system performance and reduced latency.
  • Technical Skills:
  • Data Processing: Spark with Scala, Qubole cluster management on AWS
  • Data Analytics: Presto
  • Data Visualization: Tableau, power BI
  • Data Quality: EG DQ Framework
  • Cloud Services: AWS (EMR, S3)
Data EngineeringAmazon Web Services (AWS)Extract, Transform, Load (ETL)Apache SparkData ArchitectureTableau

Meesho

Data Engineer

Nov 2021Apr 2023 · 1 yr 5 mos · Bangalore Urban, Karnataka, India

  • Data engineer at Meesho Data intelligence team
  • Description:
  • Data platform team provides a platform for business analysts and product analysts to perform analysis on critical
  • business decisions.
  • We get the event data from the application through mixpanel, the events data has been transformed to the data
  • models and are scheduled at different intervals through the day on presto cluster powered by starburst.
  • Presto cluster is a shared auto scaled cluster with ~100 node r5.12xlarge EC2 instances.
  • The data in the data models will be used by the visualization tools metabase on presto.
  • As the presto cluster is very high in terms of the cost, we migrated the data models to databricks-sql spark interface.
  • This reduced the cost per model by more than 50 % and run time improvement by around 30 %.
  • Zeppelin has been used by the business analysts to run interactive queries om presto that has been migrated to
  • spark.
  • Roles & Responsibilities in Development:
  •  Preparing technical design documents.
  •  Development, Unit testing, Code review.
  •  Responsible for scoping the amount of work, prioritizing and taking bottom line for the team
  • deliverables.
  •  Converting business requirements to functional specifications.
  •  Platformi-sing of data bricks sql using prism framework.
  •  Lead the team and completed the migration of models to the spark.
  •  The migration of 500 models from presto to databricks sql, reduced the cost by more than 50 % and performance
  • by 30 %.
  •  Helped the users (BA’s and PA’s) with the transition from presto to spark with the documentation and training
  • required
  •  Interface with the customers to ensure expectations are met.
  •  Assist in KT and enablement sessions within the team both on existing applications and new technology.
  • Ensure that the tech/product/application/domain skills are kept updated regularly.
  •  Tracking data lineage using datahub, Meta data integration (Hive metastore) and centralization.
Data EngineeringScalaPrestoAmazon Web Services (AWS)Extract, Transform, Load (ETL)Microsoft Azure+2

Walmart labs india

Data Engineer

Jul 2019Nov 2021 · 2 yrs 4 mos · bangalore

  • DESCRIPTION:
  • This project mainly deals with pricing systems for Wal-Mart stores/Ecommerce to generate the price gaps
  • between Walmart and its competitors using external third party data (RDS, Nielsen and CIA) for U.S.A. The
  • project is to build a data pipeline for moving data between RDBMS systems, Cassandra to cloud Hadoop
  • environment for data processing to the acceleration layer and then data will be processed to the consumption
  • layer SAP Hana, Apache Druid. From the consumption layer the reports will be generated on Looker(Druid data),
  • Price performance (SAP Hana) for Wal-Mart Stores/Ecom.
  • The main objective of the project is to generate the price gaps between the Walmart and its competitors
  • to help the buyers to assist on the pricing with less manual pricing decisions.
  • This includes the Data Quality Framework(DQ), Data lineage being tracked using Apache Spline.
Data EngineeringScalaPrestoAmazon Web Services (AWS)Extract, Transform, Load (ETL)Apache Spark+2

Sungard - now part of fis

Senior Developer

Mar 2018Jul 2019 · 1 yr 4 mos · Pune Area, India

Infosys

Senior systems engineer

Dec 2014Mar 2018 · 3 yrs 3 mos · Bengaluru Area, India

  • Having 3 years of experience in IT application development on Hadoop technologies.
  • Has exposure to analysis, design, coding, testing and maintenance.
  • Domain knowledge is primarily on Retail and Banking domain.
  • Extensive experience in Big data technologies –HDFS, MapReduce, PIG, HIVE, Sqoop, core Java, Unix environment.
  • Ability to work independently and support the application.
  • An effective leader with proven abilities in working with cross-cultural teams, guiding members and enabling knowledge sharing among the team.
  • Excellent analytical, communication, Leadership qualities and interpersonal skills.
  • Quick learner and ability to adapt to dynamic work environments.

Education

Sri Krishnadevaraya University

Bachelor of Technology (BTech)

Jan 2010Jan 2014

Stackforce found 100+ more professionals with Data Engineering & Apache Spark

Explore similar profiles based on matching skills and experience