Arvind Kumar

Data Engineer

Hyderabad, Telangana, India3 yrs 9 mos experience
Highly Stable

Key Highlights

  • Expert in building scalable ETL pipelines.
  • Strong background in data engineering and cloud technologies.
  • Proficient in optimizing data workflows for efficiency.
Stackforce AI infers this person is a Data Engineer specializing in scalable ETL solutions for enterprise-level data processing.

Contact

Skills

Core Skills

Data EngineeringApache SparkEtlBackend DevelopmentApi Development

Other Skills

PythonSQLNoSQLDatabricks ProductsHadoopDatabricksCore JavaRelational DatabasesJenkinsETL TestingLarge Scale DevelopmentDABAlgorithmsSoftware DevelopmentData Pipelines

About

โ–  ๐——๐—ฎ๐˜๐—ฎ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ | ๐—ก๐—œ๐—ง ๐—ง๐—ฟ๐—ถ๐—ฐ๐—ต๐˜† | ๐—”๐——๐—ฃ ๐—ฃ๐˜ƒ๐˜ ๐—Ÿ๐˜๐—ฑ โ˜… Designing scalable data pipelines & analytics-ready systems at enterprise scale. โ˜… Iโ€™m a Data Engineer at ADP Pvt Ltd, focused on building robust, high-performance data pipelines that transform raw enterprise data into business-ready insights. โ˜… My work revolves around real production data, complex source systems, and scalable platforms โ€” not toy projects or theory. (Background includes Java backend development, which strengthens my system design and data pipeline architecture.) ๐Ÿ’ผ ๐—ช๐—ต๐—ฎ๐˜ ๐—œ ๐——๐—ผ ๐—ฎ๐˜€ ๐—ฎ ๐——๐—ฎ๐˜๐—ฎ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ โ˜… Design and build end-to-end ETL pipelines on Databricks using Python & PySpark Process and transform data from enterprise Systems of Record (SORs) โ˜… Optimize Spark jobs for performance, scalability, and cost Model and store analytics-ready data in MongoDB & cloud data platforms โ˜… Write efficient SQL for validation, transformation, and analytics Monitor, debug, and analyze pipelines using Splunk โ˜… Collaborate with stakeholders to deliver reliable, business-ready datasets ๐Ÿ”ฅ๐—ช๐—ต๐—ฎ๐˜ ๐——๐—ถ๐—ณ๐—ณ๐—ฒ๐—ฟ๐—ฒ๐—ป๐˜๐—ถ๐—ฎ๐˜๐—ฒ๐˜€ ๐— ๐—ฒ โœ” Strong fundamentals in distributed data processing โœ” Hands-on experience with enterprise-scale data systems โœ” Focus on performance, data quality, and reliability โœ” Practical understanding of how data platforms run in real companies ๐Ÿ”” ๐—ช๐—ต๐˜† ๐—™๐—ผ๐—น๐—น๐—ผ๐˜„ ๐— ๐—ฒ ๐—™๐—ผ๐—น๐—น๐—ผ๐˜„ ๐—บ๐—ฒ ๐—ถ๐—ณ ๐˜†๐—ผ๐˜‚ ๐˜„๐—ฎ๐—ป๐˜: Practical Databricks & PySpark concepts used in production Real insights into enterprise data engineering Career guidance for data engineers (not surface-level content) Learn how to build scalable & reliable data pipelines ๐Ÿ‘‰ Follow for real-world data engineering insights ๐Ÿ› ๏ธ ๐—–๐—ผ๐—ฟ๐—ฒ ๐——๐—ฎ๐˜๐—ฎ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด ๐—ฆ๐˜๐—ฎ๐—ฐ๐—ธ Python | PySpark | Databricks | AWS | SQL | MongoDB | Kafka | Splunk | Snowflake | Data Modeling | Java

Experience

Sigmoid

Senior Data Engineer

Mar 2026 โ€“ Present ยท 0 mo ยท Bengaluru ยท On-site

Data EngineeringApache Spark

Adp

2 roles

Data Engineer

Promoted

Oct 2022 โ€“ Mar 2026 ยท 3 yrs 5 mos ยท On-site

  • Dynamic and result-oriented data engineer with 3+ years of experience in data engineering and cloud technologies
  • Specializes in Python, SQL, and NoSQL databases
  • Proven expertise in building scalable, high-performance ETL pipelines and optimizing data workflows for better efficiency
  • Transitioned from a mechanical engineering background to a strong career in data engineering, demonstrating adaptability and a sharp learning curve
  • Extensive experience with ETL processes using Databricks and Apache Spark to handle large-scale data processing
  • Hands-on experience with cloud platforms, particularly AWS, for data storage and processing
  • Designed and managed data pipelines for storing and retrieving data from MongoDB and other NoSQL databases
  • Proficient in using Splunk for monitoring and analyzing data logs
PythonSQLNoSQLDatabricks ProductsHadoopData Engineering+1

Back End Developer

Jul 2022 โ€“ Oct 2022 ยท 3 mos ยท On-site

  • Developed backend components using Core Java to process and expose structured data via REST APIs.
  • Integrated Splunk with application logs to create custom dashboards for real time monitoring and alerting .
  • Created summary index for accessing the data for long period .
  • Exposure to performance tunning and structured logging -valuable in real-time data environment.
Core JavaRelational DatabasesJenkinsBackend DevelopmentAPI Development

Education

National Institute of Technology, Tiruchirappalli

Bachelor of Technology - BTech

May 2018 โ€“ May 2022

Stackforce found 100+ more professionals with Data Engineering & Apache Spark

Explore similar profiles based on matching skills and experience