Prince Singh

Software Engineer

Bengaluru, Karnataka, India13 yrs 2 mos experience
Most Likely To Switch

Key Highlights

  • Over 12 years of experience in Big Data.
  • Expert in Apache Spark and Hadoop technologies.
  • Proven track record in delivering data-driven solutions.
Stackforce AI infers this person is a Big Data Engineer with expertise in Telecom and Travel industries.

Contact

Skills

Core Skills

Apache SparkHadoopHive

Other Skills

SparkScalaOozieSpark StreamingFlumeKafkaCassandraMySQLSpark GraphRest APIsBigQueryGCP

About

I have 12+ years of experience in BigData. As a BigData Developer i have worked for many domains which includes Banking,Travel , Digital Marketing and Telecom. I have always focused on understanding the needs of the end user and delivering the desired product to them.Being a technology enthusiast I have always kept myself updated with the latest trends in the technologies and tool which are currently adopted by the industries. Technical Skills :- Spark Graph,Spark Scala, Hive, Hadoop, Impala, Presto, Oozie, Hue, Sqoop, Hbase , Airflow, Databricks , Rest Api's , Spark Streaming, Bigquery, GCP

Experience

13 yrs 2 mos
Total Experience
2 yrs 2 mos
Average Tenure
2 yrs 7 mos
Current Experience

Wayfair

Lead Data Engineer

Oct 2023Present · 2 yrs 7 mos · Bengaluru, Karnataka, India · Hybrid

Adobe

Senior Data Engineer

Jun 2021Oct 2023 · 2 yrs 4 mos · Bengaluru, Karnataka, India

Absolutdata analytics-an infogain company | adobe

Senior Data Engineer

Feb 2019Jun 2021 · 2 yrs 4 mos · Bengaluru, Karnataka, India

Otsi - object technology solutions india pvt. ltd

Senior Software Engineer

Jan 2018Jan 2019 · 1 yr · Bengaluru Area, India

  • Project : Smart Profile Server (HPE R&D)
  • Smart Profile Server Data Analytics Layer is an application which aims at enabling service providers to analyze collected network data including internet traffic, and deliver results that can be used to build and expose an enriched profile to the Service Provider’s 3rd party partners or internal departments. The Smart Profile Server Data Analytics Layer is designed to host different analytics solutions bundled into Value Packs that are consistent at business management level.
  • Tenure: January, 2018 – Present
  • Responsibilities:
  • Creating a four node HDP(2.5.3 ) cluster for spark development.
  • Creating spark driver program in scala .
  • Creating a framework for spark in scala to load the data , do transformations and then sink the data .
  • Processing the data using Spark Streaming with different window n slide interval.
  • Creating lambda architecture in scala to merge streaming data and referential data.
SparkScalaHadoopHiveOozieSpark Streaming+1

Harman international india pvt. ltd.

Senior Software Engineer

Jul 2016Jan 2018 · 1 yr 6 mos · Bengaluru Area, India

  • Project : Group Marketing Platform(GMP), TUI Travel, UK
  • GMP is a HDFS based Data Warehousing platform built for TUI Travel, UK. This Big Data Lake has been built to enable Campaign Management and Analytics for TUI Travel. The solution handles upstream data coming in from different data sources, in formats like CSV, JSON,XML through flume, Kafka, FTP Feed Files. The source data is ingested into HDFS in the AVRO format. Transformation and Cleansing is applied and the ingested data is standardized and made available in Impala/Parquet Layer and then pushed into MS SQL Server Database for consumption by various Campaign Management Tools. Tenure: 15thJuly,2017 – Till January 2018
  • Responsibilities:
  • Getting data from one of the source i.e TUI webserver using flume and kafka.
  • Conversion of data into avro.
  • Using spark for processing data after receiving it in our hadoop cluster.
  • Performing transformations.
  • Pushing the data to mysql for campaigning.
  • Project : Interim Analytics, TUI Travel, UK
  • Interim Analytics is a HDFS based Data Warehousing platform built for TUI Travel, UK. This Big Data Lake has been built to enable a centralized platform for all the source of the TUI travel and push the data from GMP cluster to Interim analytics cluster on daily basis to carry on further analysis.
  • Tenure: 16th Oct ,2016 – 15th July,2017
  • Responsibilities:
  • Worked on an Oozie based solution for building Analytical Data Lake that mirrors the GMP Big Data Lake.
  • End-to-end migration from different sources to GMP Big Data Lake platform.
  • Scheduling the movement of data from one cluster to other using Oozie.
  • Performing transformations using hive.
  • Project : Customer Satisfaction Questionnaire (CSQ), TUI Travel, UK
  • CSQ is a HDFS based Data Warehousing platform built for TUI Travel, UK. This Big Data Lake has been built to store the customer feedback based on their travelling experience. The CSQ data is received in json format and then cleansed and transformed using hive.
HadoopSparkFlumeKafkaHiveApache Spark

Wipro

Project Engineer

Jan 2013Jun 2016 · 3 yrs 5 mos · Bengaluru Area, India

  • Project : Customer Journey Analysis for CapitalOne(US)
  • A customer journey map for Capital One is a framework that enables BI agents to improve customer experience. It documents the customer experience through their perspective, helping best understand how customers are interacting and helps identify areas for improvement moving forward. The "Journey" Project was rooted in data driven research and visually represents the different phases of customers experience based on a variety of dimensions such as sentiment, goals, touch points, and more.
  • Tenure: 29th April,2015 – April 2016
  • Responsibilities:
  • Filling the Hadoop data lake (HDFS) from various sources like Teradata and Mainframes.
  • Involved in refinery work to do the transformation in Scalding.
  • Involved in creating hive tables.
  • Involved in Cassandra Data Modelling.
  • Bulk load and streaming data from HDFS to Cassandra through SPARK
  • Project : Cerberus
  • This is an internal Wipro project which is used for fraud detection and analysis. It is an initiative from the CEO of the Wipro which aims at detecting the employee’s within the campus who are involved in fraud whether related to their attendance, password sharing and also if there is some fault related to payment etc and many more. After getting a successful response from the organization ,now it is being presented to other organization as a product for service.
  • Tenure: 20th May,2014 – 14th Feb,2015
  • Responsibilities:
  • Development of all the hive queries required for developing the rules for fraud detection.
  • Using Cloudera platform for hadoop.
  • Transferring of the output from hive to mysql using sqoop, so that it can be used for report generation.
  • Automation of queries using shell scripting to run on scheduled basis.
HadoopHiveCassandraSparkApache Spark

Education

AMITY UNIVERSITY

B.Tech — CS&E

Jan 2008Jan 2012

D.A.V PUBLIC SCHOOL

12th — PCM

Jan 2008Jan 2008

D.A.V PUBLIC SCHOOL

10th

Jan 2006Jan 2006

Stackforce found 100+ more professionals with Apache Spark & Hadoop

Explore similar profiles based on matching skills and experience