Jacob Francis Paruvummel

Data Engineer

Bengaluru, Karnataka, India5 yrs 2 mos experience
Highly Stable

Key Highlights

  • 4+ years of experience in data engineering.
  • Expert in building scalable ETL pipelines.
  • Proven ability to optimize data processing efficiency.
Stackforce AI infers this person is a Data Engineer specializing in SaaS solutions with a focus on ETL and data integration.

Contact

Skills

Core Skills

Etl PipelinesDatabricksData Migration

Other Skills

Active LearningAirflowAzure Data FactoryAzure Data LakeAzure DatabricksAzure DevOps ServerBig DataDelta LakeEnglishInterpersonal SkillsJSONJavaKafkaLeadershipLifelong Learning

About

Data engineer with an experience of 4+ years who have demonstrated history of designing and implementing high-quality, scalable data pipelines for enterprise clients. Deep knowledge in Python with PySpark, as well as cloud platforms like Azure. Proven ability to improve data processing speed and efficiency, while also optimizing operational workflows. Experienced in building ETL Pipelines for data ingestion, data extraction, reconciliation, and data migration, ensuring data integrity and consistency. Collaborative team player with experience in agile environments, adept at driving business value through data-driven insights.

Experience

Wesco

Data Engineer

May 2024Present · 1 yr 10 mos · Bengaluru, Karnataka, India · Hybrid

Infosys

4 roles

Technology Analyst

Oct 2023May 2024 · 7 mos

  • Delivered a cutting-edge, low-code ETL pipeline from the ground up to optimize O9 data integration on Databricks,
  • leveraging Python and PySpark to automate key processes such as data extraction, pre-validation, data migration, and
  • reconciliation.
Python (Programming Language)PySparkDatabricksETL Pipelines

Senior System Engineer

Promoted

Oct 2022Oct 2023 · 1 yr

  • Spearheaded the creation of robust ETL pipelines utilizing Databricks, seamlessly integrating data from disparate sources
  • and enabling the successful execution of the DC Split initiative.
  • Streamlined data transfer from Databricks to SAS, we integrated SAS application with Databricks and Airflow Astronomer
  • DAGs. This eliminated manual batch runs, leading to improved efficiency, reduced manual effort by 95 percent and better
  • data-driven decisions.
  • Optimized the data extraction from Oracle DAS to Delta Lake and Kafka using Databricks and Python, resulting in in-
  • creased efficiency by 50 percent and improved data accuracy (leadership, team building, project management, problem
  • solving)
  • Adroitly designed and executed an ASN integration pipeline in Databricks that facilitates seamless data migration, JSON
  • message parsing, custom validation, Kafka publishing, and rapid, precise data delivery.
  • Orchestrated a seamless integration of Databricks, Pyspark and Selenium to craft a self-reliant, sophisticated dashboard
  • and report download system that provides stakeholders with precise, timely reconciliation reports, improving business
  • visibility and reducing manual workload, human error by 80 percent
DatabricksSASAirflowPythonDelta LakeKafka+1

System Engineer

May 2021Oct 2022 · 1 yr 5 mos

  • Collaborated with senior team members to support various data-related tasks and projects including sales reconciliation
  • and integration batch monitoring
  • Assisted in troubleshooting and resolving data-related issues, maintaining data integrity.
  • Contributed to system improvements and optimizations, enhancing overall performance for EDI850 integrations by 60
  • percent.

System Engineer Trainee

Jan 2021May 2021 · 4 mos

  • • Gained foundational knowledge of Java, React, Springboot

Education

S.C.M.S. School of Engineering Technology, Karukutty P.O., Ernakulam

Bachelor of Technology — Civil Engineering

Jan 2016Jan 2020

Stackforce found 100+ more professionals with Etl Pipelines & Databricks

Explore similar profiles based on matching skills and experience