Kamesh Bhallamudi

Software Engineer

Bengaluru, Karnataka, India16 yrs experience
Most Likely To SwitchHighly Stable

Key Highlights

  • Expert in building scalable data ingestion platforms.
  • Contributed to major open source projects like Apache Drill.
  • Proven track record in AWS data migration.
Stackforce AI infers this person is a Data Engineering expert with significant contributions to cloud and open source technologies.

Contact

Skills

Core Skills

Data EngineeringDistributed SystemsCloud EngineeringOpen Source Development

Other Skills

ANSI SQLAWSAlgorithmsAnalyticsApache SparkAvroCDCCapacity planningClickstream ingestionCore JavaData Infrastructure monitoringData MiningData StructuresData auditingData cleansing

About

Experience in Software Development with exposure to - Data Ingestion - Distributed Systems - MicroServices Apart from this, I am interested in contributing to open source projects and developing tools using latest technologies.Projects I have been contributing includes : - Apache Hadoop MapReduce - Apache Hadoop YARN - Apache Drill Tools I have developed includes : - hadoop-qctree Technologies that I work on: - Hadoop, HBase, Hive, MongoDB, Spark, AWS, Kafka, Airflow, MySQL - Java, Scala, Python - MicroServices (Springboot) Github : https://github.com/kameshb My blog on Drill Mongo Storage Plugin

Experience

Uber

2 roles

Staff Software Engineer

Promoted

Sep 2022Present · 3 yrs 6 mos

Senior Software Engineer

Oct 2020Aug 2022 · 1 yr 10 mos

  • Working on building a scalable Competitive Intelligence Data platform. Responsible for
  • 1. Data ingestion, data quality, and data excellence
  • 2. Data Infrastructure/Resources monitoring & optimization
  • 3. Designing and developing a scalable Matching platform
Data ingestionData qualityData excellenceData Infrastructure monitoringData optimizationScalable platform design+2

Intuit

3 roles

Staff Data Engineer

Promoted

Aug 2018Oct 2020 · 2 yrs 2 mos · Bengaluru Area, India

  • AWS Migration
  • AWS migration is to migrate data & ingestion pipelines to AWS infrastructure. I was responsible for
  • 1. Designing strategies for the data & pipelines migration
  • 2. Capacity/Infrastructure planning
  • 3. Security review
  • 4. Leading the development effort for migration
  • Migrated around 8000+ tables, 15PB+ data without any downtime.
  • Unified Ingestion Platform
  • Unified Ingestion platform is a data ingestion platform that replicates data from any source type (DB, File, API, etc) to Data lake (AWS) in near real-time. I was responsible for
  • 1. Designing and developing Kafka connect based collator service
  • 2. Designed and developed an efficient data reconciliation platform. Some of the features include data snapshotting, backfill, replayability, etc
  • 3. Designed and led the development of Self-serve onboarding of datasets/tables. It supports provisioning of hardware resources/components for datasets, efficient scheduling of pipelines along with pipeline setup.
  • 4. Infrastructure/resources/cost optimization
AWSData migrationData ingestionKafkaInfrastructure planningSecurity review+2

Senior Data Engineer

Promoted

Feb 2016Jul 2018 · 2 yrs 5 mos · Bengaluru Area, India

  • R3D3
  • I was part of the R3D3 team. It collects CDC (change data capture) events (events from Redo logs/Bin logs) in real-time from RDBMS and replicates them to Hive in batches. I was responsible for
  • 1. Designing and developing collator service, which replicates data from Kafka and stores it in HDFS.
  • 2. Designed and contributed to efficient Data Merging/reconciliation algorithms.
  • 3. Self serve platform for onboard datasets
  • 4. Capacity planning and resources optimization
  • EventTrackingSystem
  • Event tracking system is a data pipeline auditing framework. I have designed and led the development of this system. This framework detects if there is any event loss in the pipeline in near real-time.
CDCKafkaHDFSData mergingCapacity planningData Engineering+1

Data Engineer

Apr 2014Jan 2016 · 1 yr 9 mos · Bengaluru Area, India

  • Profile Matching Framework
  • Entity/Profile matching framework is a scalable matching framework that does the entity/profile matching in real-time and develops relationships between entities in a Graph Store. Developed matching algorithms based on fuzzy matching techniques.
  • Profile Attributes Aggregator
  • This framework collects attributes of a profile from various different sources, unions the different attributes to create a unified profile of an entity, and stores it in Profile Store.
Entity matchingFuzzy matchingProfile aggregationData Engineering

The apache software foundation

Apache Drill Committer

Oct 2014Dec 2018 · 4 yrs 2 mos

  • Apache Drill is distributed query engine on top of any data store.
  • 1. Developed Mongo Storage Plugin for Apache Drill. This capability allows Apache Drill users to query MongoDB using ANSI SQL.
  • 2. Developed Kafka Storage Plugin for Apache Drill. This capability allows Apache Drill users to query Kafka using ANSI SQL.
  • 3. Enhanced Avro file format support for Apache Drill. Added support for Schema validation, support for nested data structures.
MongoDBKafkaAvroSchema validationOpen Source Development

Homeshop18

Software Development Engineer

Aug 2012Mar 2014 · 1 yr 7 mos · Bengaluru Area, India

  • Clickstream Ingestion Platform
  • Designed and developed a Clickstream ingestion Platform. This platform
  • 1. Ingests clickstream and transaction logs from different App servers to Hadoop.
  • 2. Standardizes and cleanses clickstream and transaction logs and stores them in Hive Warehouse
  • Analytics
  • Developed applications for
  • 1. Finding the Clickthrough and Conversion rate of each web page.
  • 2. Cart abandonment.
  • 3. Top Selling products/queries
  • Search & Autosuggestion
  • I was part of the search team developing/analyzing users' queries.
  • 1. Developed applications to normalize user queries,
  • 2. Mining frequently searched queries in a user session
  • 3. Developing query recommendation using the collaborative technique
  • Recommender System
  • Designed and developed Recommender systems
  • 1. Product recommendation using content & collaborative filtering
  • 2. Personal recommendation using Matrix Factorization Techniques.
  • 3. Developed scalable Recommendentation API
Clickstream ingestionData cleansingAnalyticsData Engineering

Pramati technologies

Senior Development Engineer

Jul 2011Jul 2012 · 1 yr · Hyderabad Area, India

  • Developed Video analytics applications for TubeMogul client using Hadoop.

The apache software foundation

Apache Yarn/MapReduce Contributor

Jul 2010Jun 2013 · 2 yrs 11 mos

  • Contributed around 16 patches, all of which are accepted/committed.
  • Contributions link - tinyurl.com/ypeackdz

Huawei technologies india pvt ltd

Software Engineer

Jan 2010Jun 2011 · 1 yr 5 mos · Bengaluru Area, India

  • I was part of the Platform team, worked on open source technologies like Apache Hadoop MapReduce/Yarn and Restlet. As part of MapReduce/Yarn enhancement & stabilization, fixed many critical bugs and contributed many new features. Actively participated in Apache Open Source community and contributed some of the enhancements done back to open source. I have contributed 16 patches and all of which are accepted.
  • Contributions - tinyurl.com/ypeackdz

Education

Indian Institute of Technology, Kharagpur

M.Tech — Computer Science and Data Processing

Jan 2007Jan 2009

Andhra University

M.Sc — Mathematics

Jan 2005Jan 2007

Stackforce found 100+ more professionals with Data Engineering & Distributed Systems

Explore similar profiles based on matching skills and experience