A

AK J.

Data Engineer

Frisco, Texas, United States17 yrs 9 mos experience
Most Likely To SwitchHighly Stable

Key Highlights

  • 14 years of experience in data engineering.
  • Expert in AWS, GCP, and Azure technologies.
  • Pivotal in building scalable data solutions.
Stackforce AI infers this person is a Cloud Computing Data Engineer with extensive experience in building scalable data solutions.

Contact

Skills

Core Skills

Data EngineeringCloud Computing

Other Skills

AWSAWS CWIAWS CloudFormationAWS Command Line Interface (CLI)AWS GlueAWS Identity and Access Management (AWS IAM)AWS LambdaAWS SageMakerAWS Step FunctionsAirflowAmazon AthenaAmazon CloudFrontAmazon DynamoDBAmazon EC2Amazon QuickSight

About

I am a Senior Data Engineer at Amazon Web Services (AWS), with over 14 years of experience in designing and developing data warehouse and business intelligence solutions using AWS, GCP, Azure, Data Bricks, Snowflake, Python, Pyspark,Informatica, Oracle, and Unix. I have a master's degree in computer science and a bachelor's degree in electrical and electronics engineering. At AWS, I am pivotal in designing and building the Demand Planning Tool (DPT) project, which deals with private pricing and financials of AWS accounts and their incentives based on their usage over the contract period. My extensive experience in developing scalable data solutions, along with my ability to thrive in uncertain environments and my strong sense of curiosity, makes me an ideal fit for this role. My proficiency in technologies such as AWS, Azure, Spark, GCP combined with my expertise in Python, and SQL, aligns well for any requirement in Data Engineering.

Experience

Amazon web services (aws)

Senior Lead Data Engineer (L5)

Apr 2020Present · 5 yrs 11 mos · Dallas, Texas, United States · Remote

  • 𝓟𝓻𝓸𝓳𝓮𝓬𝓽: 𝓓𝓟𝓣(𝓓𝓮𝓶𝓪𝓷𝓭 𝓟𝓵𝓪𝓷𝓷𝓲𝓷𝓰 𝓣𝓸𝓸𝓵)
  • 🔹🔹I was pivotal in designing and building the DPT Project. Project mainly deal with Private pricing; financials of revenue relate to AWS accounts (Both Private and Gov.). 🔹🔹
  • ✦ Interacted with different stakeholders, Product leaders to get business requirements and
  • architected the end-to-end design, write efficient real-time and batch processing data pipelines to
  • provide data needed for analytics.
  • ✦ Developed Spark UDFs to reduce the SQL complexity, maintain data consistency and
  • improve the code reusability.
  • ✦ Responsible for building the CloudFormation templates for SNS, SQS, Elastic Search, Dynamo DB,
  • Lambda, EC3, IAM, Cloud Watch Services
  • ✦ Collaborated with cross functional teams to come up with end-to-end scalable and reliable data
  • solutions for successful product/feature releases.
  • ✦ Created critical goal dashboards using visualization tools, so the product teams can access and
  • monitor key metrics.
  • ✦ Built end-to-end data pipelines, including data ingestion, transformation, and loading processes
  • mostly written in Python or Py Spark.
  • ✦ Data Extraction, aggregations and consolidations of Adobe data, Sales data, Finance data with AWS
  • Glue using PySpark.
  • ✦ Created multiple lambda functions and automated the deployment and configured it to receive
  • events from s3 bucket for file events such as insert/delete/updates.
  • ✦ Developed Py Spark code for hundreds of jobs using AWS Glue and EMR architecture.
  • ✦ Good Hands-on Experience with AWS Services like S3, EC2, IAM, Data pipelines like AWS Step Functions, Lambda, Data pipelines, Glue, Terraform, VPC, Redshift, Athena, Dynamo DB etc.
  • 𝑬𝒏𝒗𝒊𝒓𝒐𝒏𝒎𝒆𝒏𝒕: SQL, Python, MySQL, Py Spark, AWS (EC2, S3, EMR, Redshift, Athena, Glue, SNS, SQS, VPC, Cloud Formation, Cloud Watch), Bitbucket, Airflow, SFDC.
AWSPythonPySparkSQLAWS GlueAWS Lambda+11

Google

Senior Data Engineer

Jan 2018Apr 2020 · 2 yrs 3 mos · Menlo Park, California, United States · On-site

  • 𝓟𝓻𝓸𝓳𝓮𝓬𝓽 : 𝓡𝓸𝓲𝓷𝓷
  • Roinn is a Datamart which provides a consolidated, reliable, and easy-to-query view of billing data. It focuses on the Accounts Receivable (AR) position, and a customer dimension table, for use primarily by accounting and finance Operations.
  • 𝓟𝓻𝓸𝓳𝓮𝓬𝓽 : 𝓣𝓱𝓻𝓮𝓼𝓱𝓸𝓵𝓭 𝓒𝓸𝓵𝓵𝓮𝓬𝓽𝓲𝓸𝓷𝓼
  • Threshold Billing customers are typically invoiced for small amounts (circa $500). When they do not pay, the debts are placed with external agencies (OCAs) for collection. The Threshold Collections data drives this process, both in terms of generating the monthly placements and tracking repayments.
  • 𝓟𝓻𝓸𝓳𝓮𝓬𝓽 : 𝓖𝓸𝓸𝓰𝓵𝓮 𝓟𝓵𝓪𝔂/𝓣𝓪𝓵𝓸𝓼
  • Finance BI&A support Play Finance by producing the Talos DataMart. This provides reporting on Financials, Users, Apps, Partners, and Ads on Play to support the Play business.
  • 𝑬𝒏𝒗𝒊𝒓𝒐𝒏𝒎𝒆𝒏𝒕: Oracle r12, Google Technologies: Plx, F1, Dremel, CNS, Placer, Borg, Google3, Rapid, Stubby, g3doc, Pentaho, Python, GCP, Big query, Py spark, Gcs Bucket, G-Cloud Function, sql, etc.
PythonGoogle Cloud Platform (GCP)SQLInformaticaPentahoData Engineering+1

Gilead sciences

Sr Data Engineer

Jun 2016Dec 2017 · 1 yr 6 mos · Foster City, California, United States · On-site

  • Worked with business team to gather requirements for reporting needs.
  • Prepared ETL standards, Naming conventions and wrote ETL flow documentation for Stage, ODS and Mart.
  • Responsible for collecting information regarding the design on how to move the data for each system to decide on what will be needed to be taken to do the archival for each system.
  • Worked with the subject-matter expert to get the increment logics of all the systems on a daily basis
  • Provided Support and effort for process optimization and ETL automation.

Csaa insurance group, a aaa insurer

ETL Data Engineer.

Jan 2015Jun 2016 · 1 yr 5 mos · Glendale,AZ · On-site

  • Estimated project levels of effort and resource requirements and working with staff to understand scope of effort.!!
  • Managed project team of 8-10 members in an onshore-offshore model.
  • Leaded and guided development of an Informatica based ETL architecture for Insurance Systems.
  • Planned and executed schedules for major and minor releases and other technical deliverable's as well as ad-hoc support.
  • Facilitated scrum ceremonies (grooming, sprint planning, retrospectives, daily stand-ups, etc.)
  • Implemented an Informatica based ETL solution fulfilling stringent performance requirements.
  • Collaborated with product development teams and senior designers to develop architectural requirements to ensure client satisfaction with product.
  • Managed build phase and quality assure code to ensure fulfilling requirements and adhering to AAA ETL standards.
  • Supported design, development and operations in cross tracks applications for master data management, business intelligence as well Other Teams.!
  • Served as a member of the Cross track meetings within larger group, representing MDM Solutions Technical Lead to view & address any impact in MDM Solutions.
  • Leaded Unit and Integration Testing activities and assisted in User Acceptance Testing.

Genentech

ETL Lead

Jan 2013Jan 2015 · 2 yrs · San Francisco Bay Area

  • Worked as ETL Salesforce Lead & Lead Developer
  • As a ETL lead and responsible for the following activities.
  • Involve in requirement gathering session with business users.
  • Analyze the requirement and prepare Technical Specification doc.
  • Implement/manage ETL development activities with offshore team by providing them
  • sufficient guidance to develop code with quality and reusable standards.
  • SQA Support by running data loads, bug fixes during testing in SQA environment.
  • Participate in UAT sessions with business users to support data validation in UAT
  • environment.
  • Responsible for creating various project related docs like Detail design doc, STM's (Source to
  • Target Mappings), Deployment docs, Operation support docs.

Leapfrog

Sr.ETL Developer

Jan 2009Dec 2012 · 3 yrs 11 mos · Emeryville, California, United States · On-site

  • Working as Sr.ETL Developer Involved in EDW Design at Leapfrog. Currently Working as Informatica Production Support Role,Informatica Admin and Developer.

Vmware

ETL Migration Developer

Apr 2008Dec 2009 · 1 yr 8 mos · Palo Alto, California, United States · On-site

  • Involved in analyzing, defining, and documenting data requirements by interacting with the client and
  • Salesforce team for the Salesforce CRM objects.
  • ♦ Developed Mapping Document to present to Business and sales team for the Informatica process for the
  • Salesforce objects migration.
  • ♦ As a lead member of ETL Team, responsible for analyzing, designing and developing ETL strategies and
  • processes, writing ETL specifications for developer, ETL and Informatica development, administration.
  • and mentoring.
  • ♦ Performed some of the roles of Salesforce.com Administrator in the organization.
  • ♦ Worked with various salesforce.com objects like Leads, Lead Opportunities Update, Parent
  • Opportunities Update, Opportunity Record Type Update, Opportunities, Revenue Line Items, Order
  • Logs, Split Territories, Order Mappings, Attachments, Reports, and Dashboards.
  • ♦ Involved in extracting the data from Siebel source systems into Defined Staging area based on the
  • inclusion and exclusion criteria.
  • ♦ Created and edited custom objects and custom fields in Salesforce and checked the field level
  • Securities.
  • ♦ Worked on Offshore Onshore model, held meetings on a daily basis for coordination and planning in the
  • Siebel-Salesforce CRM Migration and Integration Testing.

Education

Jawaharlal Nehru Technological University

Bachelor's degree — Electrical and Electronics Engineering

Jan 2001Jan 2005

Stackforce found 100+ more professionals with Data Engineering & Cloud Computing

Explore similar profiles based on matching skills and experience