Suman Movva

Co-Founder

Slough, England, United Kingdom4 yrs 8 mos experience

Key Highlights

  • Expert in big data technologies and data engineering.
  • Co-founder of a leading data recruitment agency.
  • Extensive experience with Fortune 500 companies.
Stackforce AI infers this person is a Big Data and Data Engineering expert with a focus on recruitment.

Contact

Skills

Core Skills

Big Data TechnologiesData EngineeringData WarehousingData ProcessingData MigrationEtl Development

Other Skills

HadoopMapReduceClouderaPythonSparkLuigiAVROHIVEIMPALAAmazon S3Amazon RedshiftPIGApache SparkSAP Data ServicesBusiness Objects

About

WHO I AM An IT expert that has worked with multiple Fortune 500 companies, I have gained vast experience in data technologies. I am passionate about data, which stems from an academic background in engineering. As I enjoy what I do so much, it feels more like I am simply playing with data all day. When I am not working with data, I will be travelling, or playing and watching cricket. WHAT I DO As one of the founders of Data Buzz, I help data-focused organisations recruit the best talent to work with their data. The UK Data Recruiting Agency we have built is based on a solid foundation with all three of our co-founders having known each other for 20 years. The experience we have in cutting edge data technologies is second to none. WHO I WORK WITH? We work with data technologies consisting of programming languages such as Scala and Python Data Analytics, with visualisation tools like Power BI Cloud Implementations, AWS, Azure and GCP. Organisations we work with: - Start-up businesses. - Businesses up to $100m revenue. - Based in the UK. - Require large volumes of data to be analysed. Our clients appreciate the value data brings to their company’s key metrics. They appreciate the talent and resources they have should be passionate about keeping up to date on data trends. Who want to; - Hire the best talent to handle their data & improve retention. - Outsource their talent scouting. - Hire talent that is in sync with the company’s long term goals. - Stop missing project deadlines and improve the quality of projects. - Ensure data is analysed and presented effectively to support business decisions. HOW I HELP We find the best data talent on behalf of companies in a world where there are not many highly skilled data professionals due to data trends and technologies evolving so quickly. To support talent retention, we ensure the talent understand and believe in your organisation's long term goals from the outset. By having the most suitable people in place first time you will avoid continuous costs of having to train new employees on the same skills. WHY IT WORKS! You do not need to invest in: - R&D in scouting the top talent. - Time maintaining relationships with individuals. - Dealing with a large volume of candidates for every position. Our experience in working with data and recruitment provides an invaluable understanding of the skills and experience required of candidates to fill any specific role. CONTACT ME Visit our website to find out more: www.databuzzltd.com

Experience

Databuzzltd

Co-Founder

Present

Mitie

Lead Data Consultant

Mar 2017Present · 9 yrs 1 mo · London, England, United Kingdom

Lseg (london stock exchange group)

Lead Data Engineer

Jul 2016Mar 2017 · 8 mos · London, England, United Kingdom

  • Design and development of software in Cloudera’s Hadoop MapReduce framework.
  • Configured and ran all MapReduce programs on large node cluster using chef etc.
  • Designed and developed incremental loads from legacy, involving several technologies like sqoop, impala, python.
  • Used luigi to automate data loading into the Hadoop Distributed File System and Python to pre-process the data.
  • Performing hands-on research and software development to explore how big data technologies such as Hadoop, Apache Spark (Scala) etc. can be leveraged to increase the scalability of company's data analysis.
  • Developed Spark D-Stream’s to process data in micro batches using Spark Streaming.
  • Used Spark for batch processing and ad-hoc analytics
  • Used AVRO Serdes to handle AVRO format data in HIVE and IMPALA.
  • Experienced in implementing Spark RDD transformations, actions to implement business analysis and Worked with Spark accumulators and broadcast variables
  • Developed scripts, UDF's using both Spark SQL and Spark Core in scala for Data Aggregation, queries and verified its performance over MR jobs.
  • Designed and implemented data staging methods and stress testing of ETL routines to make sure that they don’t break on heavy loads.
HadoopMapReduceClouderaPythonSparkLuigi+5

Fanduel

Senior Data Engineer

Apr 2015Dec 2015 · 8 mos · Edinburgh, Scotland, United Kingdom

  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
  • Design and development of software in Hadoop MapReduce framework using Amazon S3, Amazon EC2, Amazon Elastic MapReduce(EMR).
  • Configured and ran all MapReduce programs on large node cluster (Amazon EC2 spot instances) with Apache Hadoop.
  • Designed and developed a Star Schema modelled data warehouse in Amazon Redshift.
  • Designed and developed incremental loads from legacy to Redshift, involving several technologies like snoop, pig, hive, python.
  • Used luigi to automate data loading into the Hadoop Distributed File System and Python to pre-process the data.
  • Successfully ran all Hadoop MapReduce programs on Amazon Elastic MapReduce framework by using Amazon S3 for Input and Output.
  • Performing hands-on research and software development to explore how big data technologies such as Hadoop, Apache Spark (Scala), Amazon Web Services, etc. can be leveraged to increase the scalability of company's data analysis.
  • Replaced Hive programs with Spark to leverage RDD's in-order to obtain lower latency for the end-user
  • Used Spark for batch processing and ad-hoc analytics
  • Created Hive tables, loading with data and writing hive queries which will run internally in map way
  • Designed and implemented data staging methods and stress testing of ETL routines to make sure that they don’t break on heavy loads.
HadoopMapReduceAmazon S3Amazon RedshiftSparkPython+2

Betfair

Sr Data Consultant

Dec 2014Apr 2015 · 4 mos · Hammersmith and Fulham, England, United Kingdom

  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
  • Design and development of software in Hadoop MapReduce framework using Amazon S3, Amazon EC2, Amazon Elastic MapReduce(EMR).
  • Configured and ran all MapReduce programs on 20-30 node cluster (Amazon EC2 spot instances) with Apache Hadoop to handle 1000GB/sample of WebLog data.
  • Configured a 20-30 node (Amazon EC2 spot Instance) Hadoop cluster to transfer the data from Amazon S3 to HDFS and HDFS to Amazon S3 using distcp and also to direct input and output to the Hadoop MapReduce framework.
  • Successfully ran all Hadoop MapReduce programs on Amazon Elastic MapReduce framework by using Amazon S3 for Input and Output.
  • Performing hands-on research and software development to explore how big data technologies such as Hadoop, Apache Spark (Scala), Amazon Web Services, etc. can be leveraged to increase the scalability of company's data analysis.
  • Built real-time distributed data access solutions around Apache spark
  • Replaced MapReduce programs with Spark to leverage RDD's in-order to obtain lower latency for the end-user
  • Used Spark for batch processing and ad-hoc analytics
  • Good Knowledge in using Streaming based analytics using Apache Storm & Apache Spark
  • Good Knowledge in using Messaging framework like Apache Kafka or RabbitMQ
  • Designed and transmitted a RDBMS(Oracle) Database to Hive Database
  • Created Hive queries that helped data analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Used Azkaban to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
  • Created Hive tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map way
HadoopMapReduceAmazon S3SparkHivePIG+2

Centrica

Senior Consultant

Nov 2009Dec 2014 · 5 yrs 1 mo · Staines-upon-Thames, England, United Kingdom

  • During my tenure with Centrica, managed to work on several high profile data warehousing/ data migration and big data projects, using variety of tools like Hadoop, SAP Data Services, Business Objects etc.
HadoopMapReduceAmazon S3SparkHivePIG+2

Schneider electric

Sr Datawarehousing Expert

Jan 2008Nov 2009 · 1 yr 10 mos · Galway, County Galway, Ireland

  • Understand requirements, document and implement code to the specs.
  • Data Modelling the data warehouse environment using Kimball Methodology
  • Involved in the analysis, design and development of all the interface’s using Informatica Power mart tools in interface team and interfaced with all the other tracks for business related issues.
  • Developed Informatica workflows/worklets required to extract, transform and load data from several legacy systems.
  • Developed standard and re-usable mappings and mapplets using various transformations like expression, aggregator, joiner, source qualifier, router, lookup, and filter.
  • Cleansed, Audited, Profiled, Validated and Reported ETLed data through Data Quality Management tools and transforms.
  • Address, Name and custom standardizations using Data quality.
  • Managed DI Metadata using metadata management tools.
  • Job statistics reporting using metadata management tools.
  • Written Oracle Packages, Procedures, MViews and Triggers required for Processing Batch Details, Inventory/Invoice Management.
  • Was part of scripting team for shell scripts to automate and migrate data from ODS to Data warehouse.
  • Worked on creating standard and adhoc reports, dashboard using Siebel Analytics.
  • Build the Siebel Analytics repository/metadata and define/create the subject areas where reporting and analysis are centered around.
  • Designed, Created and Tested Reports in the Dashboard and adhoc reports according to the client needs using Siebel Analytics/OBIEE.
  • Designed, Developed and Tested Data security and Dashboard security in Siebel Analytics/OBIEE
  • Writing Application Logic, Coding, Unit Testing and Support for Testing Personnel and Business Users.
  • Participated in user meetings to resolve on-going implementation & performance issues.
  • Created, updated and maintained ETL technical documentation.
HadoopSAP Data ServicesBusiness ObjectsData WarehousingData Migration

Ge capital

DWH and BI Consultant

Jan 2005Sep 2007 · 2 yrs 8 mos

InformaticaOracleSiebel AnalyticsData WarehousingETL Development

Tata consultancy services

Assistant System Engineer

Nov 2004Sep 2007 · 2 yrs 10 mos · Manchester, United Kingdom and Mumbai, IndiaManchester

  • Extracted, transformed and loaded data from operational data sources into target data warehouse using BODI.
  • Implemented profiling, extracting, cleansing, transforming, integrating and loading data into Data warehouse using BODI
  • Written database Procedures, MViews and Triggers.
  • Developed workflows required to extract, transform and load data from several GE Capital Solutions legacy systems.
  • Administration and schedule DI jobs as per daily, weekly and monthly using job server.
  • Involved in requirements gathering from the business users to build the Datawarehousing Environment.
  • Installed Business objects created repository and distribute the BO Main Key across the enterprise.
  • Created users and groups and applied row level and objects level securities using Central Management Console.
  • Involved in migration and deployment of the universes and reports across multiple domains.
  • Developed complex queries, procedures, custom functions using PL/SQL
  • Resolving loops using contexts and aliases to remove the cyclic dependencies.
  • Created reports using Business Objects functionalities like queries, slice and dice, drill down, cross tab, master detail reports, formulae, etc.
  • Creating standard full client reports and Web I reports and scheduled full client reports and Web I reports with Broadcast Agent Server.

Education

Jawaharlal Nehru Technological University

Bachelor of Technology - BTech

Jan 2000Jan 2004

Stackforce found 100+ more professionals with Big Data Technologies & Data Engineering

Explore similar profiles based on matching skills and experience