Poornima Nag Ponthagani

Product Engineer

London, England, United Kingdom10 yrs 10 mos experience
Most Likely To SwitchHighly Stable

Key Highlights

  • 10+ years in data engineering and architecture
  • Accelerated insights delivery by 40% with reusable frameworks
  • Mentored engineers, increasing team delivery by 25%
Stackforce AI infers this person is a Data Engineering expert in SaaS and Cloud solutions.

Contact

Skills

Core Skills

Data ArchitectureCloud PlatformsData EngineeringData GovernancePipeline OrchestrationEtl DevelopmentData MigrationData Quality AssuranceEmbedded SystemsControl Systems

Other Skills

Amazon Elastic MapReduce (EMR)Amazon Relational Database Service (RDS)Amazon S3Amazon Web Services (AWS)Analytical SkillsApache AirflowApache PigAzure Data FactoryAzure DatabricksBatch ProcessingBig DataCCDCCOBOLCommunication

About

With 10+ years of experience in data engineering, I help organizations build scalable, cloud-native data solutions that accelerate insight delivery, improve governance and reduce operational overhead. I architect scalable batch, streaming and event-driven pipelines on AWS, Azure, and Databricks by aligning with business goals. I’ve led Data Lakehouse implementations with Delta Lake, Iceberg and governance using Unity Catalog and Azure Purview. I deploy production ready pipelines using Airflow, ADF, Argo workflow, CI/CD, Docker, and Kubernetes collaborating closely with data scientists and mentoring engineers to drive impact and best practices. 🚀 Impact highlights: ✦ Developed a reusable data product framework for pharma usecases enabling YAML-based data ingestion, transformation and modeling. This accelerated the use case delivery by 40%, reducing development time from 3 weeks to <2 weeeks and enabling faster insights. ✦ Leveraged Generative AI (Claude, GPT) to automate data summarization reducing manual analyst workload by 40% (~60 hours/month) and cutting insight delivery time by 30%. This enabled data analysts and medical researchers to shift focus toward accelerating decision-making for product and ops teams. 🚀 Technical Skills: ✦ Languages: Python, SQL, COBOL ✦ Data Tools & Frameworks : Hadoop, PySpark, Hive, PostgreSQL, DBT, Great Expectations, Kedro, Snowflake ✦ Data Architecture: Batch processing, event-driven pipelines, real-time streaming (Kafka), Data Lakehouse Designs(Delta Lake, Apache Iceberg) ✦ Pipeline Orchestration: Apache Airflow, Azure Data Factory, Databricks Workflows, Argo Workflows ✦ Data Governance & Cataloging: Databricks Unity Catalog, Azure Purview ✦ Cloud Platforms: AWS(EC2, AWS Glue, AWS Lambda, Kinesis, EMR, S3, Redshift, RDS, DynamoDB), Azure(Azure DataFactory, ADLS Gen2, Azure Synapse, Cosmos DB, Azure Databricks) ✦ DevOps: CI/CD Pipelines (CircleCI, Azure DevOps, GitHub Actions), Docker, Kubernetes ✦ Generative AI: RAG architecture, Amazon Bedrock, Anthropic Claude Models ✦ Agentic AI Platforms: Google Agentspace, AWS Agentcore ✦ Agentic AI Integration: A2A Protocol Server Development, Multi-Agent Communication Systems 🚀 Certifications and Accreditation: - AWS Certified Developer - Associate - AWS Certified Solutions Architect - Associate - Databricks Certified Associate Developer for Apache Spark 3.0 – Associate - Kubernetes Application Developer(CKAD) – Associate - Academy Accreditation - Databricks Lakehouse Fundamentals - Academy Accreditation - Generative AI Fundamentals

Experience

Quantumblack, ai by mckinsey

4 roles

Full-time parenting

Mar 2023Apr 2024 · 1 yr 1 mo

Lead Data Engineer – Data Architecture & Agentic AI Systems

Promoted

May 2022Present · 3 yrs 10 mos

  • Enabled cross-cloud agent communication by integrating Google Agentspace and AWS Agentcore, and containerizing A2A protocol services in Docker for scalable Kubernetes-based deployment within the enterprise platform.
  • Architected and delivered scalable cloud-native data platforms across Azure, AWS, and Databricks, supporting enterprise-grade AI and analytics initiatives.
  • Designed and optimised batch, streaming and event-driven pipelines using Kafka, Delta Lake, and Apache Iceberg within Data Lakehouse architectures, alongside Snowflake cloud data warehousing to ensure data reliability, low latency, and scalability.
  • Developed real-time feature engineering pipelines and reusable PySpark modules, improving ML model iteration speed by 30% and reducing redundant engineering effort.
  • Established secure data governance and lineage tracking frameworks leveraging Unity Catalog and Azure Purview, aligning with enterprise compliance standards.
  • Orchestrated complex data pipelines using Apache Airflow, Azure Data Factory, Databricks Workflows, and Argo Workflows established automated CI/CD with GitHub Actions and Azure DevOps for reliable deployment.
  • Enhanced deployment consistency and scalability through containerisation and orchestration using Docker and Kubernetes.
  • Leveraged Generative AI applications (e.g., Claude, GPT) to automate data summarisation, reducing manual processing by 40% and accelerating insight delivery by 30%.
  • Collaborated closely with 10+ data scientists and senior business stakeholders to translate analytical requirements into scalable data architecture solutions.
  • Mentored and up-skilled 5+ junior engineers through 1:1 coaching, code reviews, and knowledge-sharing sessions, leading to a 25% increase in team delivery.
PythonSQLPySparkgreat expectationskedroSnowflake+20

Senior Data Engineer

Promoted

Jun 2020May 2022 · 1 yr 11 mos

  • Developed a reusable data product for pharma clients using Kedro, PySpark, and Python enabling configurable ingestion, feature engineering and modeling via YAML - accelerating use case delivery by 40%.
  • Built scalable batch data pipelines for the aluminium manufacturing sector using Azure Data Factory to land multi-source data into ADLS Gen2, transformed data using Databricks with Delta Lake for CDC handling and leveraged Synapse Analytics for enhanced reporting and analytics; orchestrated pipelines via Databricks Workflows and automated CI/CD using Azure DevOps.
  • Implemented advanced data governance using Unity Catalog, enabling access controls, lineage tracking and audit logging to ensure compliance and data security.
  • Deployed production-grade data pipelines to Kubernetes clusters using Argo Workflows for orchestration following dev-stage validations via Databricks Workflows - enhancing reliability, scalability and observability across environments.
PythonPySparkkedroSQLAzure Data Factoryazure data lake storage- adls gen2+13

Data Engineer

Apr 2019Jun 2020 · 1 yr 2 mos

  • Built and deployed batch data pipelines using PySpark and Kedro, ingested multi-format data from AWS S3, transforming it and storing curated datasets in PostgreSQL.
  • Orchestrated workflows using Apache Airflow ensured reliability through unit and integration testing.
  • Containerized workflows with Docker and implemented automated CI/CD using GitHub Actions.
  • Collaborated with data scientists and stakeholders to develop reusable feature engineering pipelines, enabling precise HCP targeting and campaign personalization, reducing segmentation time by 30% and improving marketing response rates by 15%.
PySparkkedroApache AirflowAmazon S3Amazon Elastic MapReduce (EMR)PostgreSQL+7

Deloitte

2 roles

Data Engineer

Promoted

May 2018Nov 2018 · 6 mos · Chennai Area, India

  • Role : Data Engineer consultant
  • Project : Customer Liability .,
  • Creating Bigdata solutions to ensure that systems and payment models in PayPal are designed to make customers feel safe and defining customer liability using complex financial rules and reconciliation processes
Data PipelinesTroubleshootingETL ToolsHivePySparkUnix Shell Scritping+4

Junior Data Engineer

Feb 2018Apr 2018 · 2 mos · Chennai Area, India

  • Role: Junior Data Engineer
  • Project : Watch – Business health Modelling .,
  • Building an in-house product of PayPal – ‘WATCH’ for certifying the quality of data (trustworthiness) by predicting data behavior based on data collected from all projects across PayPal and additionally derived metrics. Any piece of data generated in any project can be monitored through ‘WATCH’ portal in PayPal, which mainly dealt with business health check based on these data
Data PipelinesTroubleshootingETL ToolsHivePySparkUnix Shell Scritping+4

Cognizant

2 roles

Bigdata Developer

Sep 2015Oct 2017 · 2 yrs 1 mo

  • Role: Bigdata Developer
  • Project : Viewership Activity Reporting.,
  • Building a data lake to provide set top box analytics data to the clients by performing various ETL operations on hourly and daily basis, to be used by downstream teams/ATT Data scientists for AD-Targeting, market/other business analysis
  • Technical Skills : Hive, Pig, Sqoop, Shell scripting and Oracle
Data PipelinesETL ToolsHiveApache PigSqoopShell Scripting+4

Informatica Developer

Nov 2014Sep 2015 · 10 mos

  • Role: Informatica Developer
  • Project : DTV Hadoop Migration .,
  • Data audit between files of DATALAKE and GRID for obtaining the accuracy of data in two systems during the process of data migration from old databases to new databases
InformaticaToadOracle DatabaseSQLShell ScriptingData Migration+1

Isro - indian space research organization

Embedded Software Developer

May 2014Oct 2014 · 5 mos · India

  • Engineered device and application for a Live project in ISRO on Temperature Control of Satellite Cooling Using Arm Microcontroller Board by passing Liquid nitrogen to the heat shield of PSLV ROCKET.
COBOLSCADAEmbedded SystemsControl Systems

Education

International Institute of Information Technology Hyderabad (IIITH)

Bachelor of Technology (B.Tech.)

Jan 2010Jan 2014

Stackforce found 100+ more professionals with Data Architecture & Cloud Platforms

Explore similar profiles based on matching skills and experience