Ajinkya Chatufale

Software Engineer

Pune, Maharashtra, India11 yrs 3 mos experience
Most Likely To SwitchHighly Stable

Key Highlights

  • 10 years of experience in Software Development Life Cycle.
  • Expertise in AWS and Big Data technologies.
  • Proficient in Data Engineering and Real-Time Data Processing.
Stackforce AI infers this person is a Data Engineer specializing in Big Data and Cloud technologies across Fintech and Healthcare industries.

Contact

Skills

Core Skills

Big DataAwsData EngineeringSoftware Development

Other Skills

AWS CloudTrailAWS DataSyncAWS GlueAWS Identity and Access Management (IAM)AWS LambdaAWS SageMakerAWS Step FunctionsAmazon CloudFrontAmazon Elastic MapReduce (EMR)Amazon RedshiftAmazon Relational Database Service (RDS)Amazon VPCAthenaBig Data AnalyticsCassandra

About

With 10 years of experience in modern phases of the Software Development Life Cycle (SDLC). * I specialize in cutting-edge technologies such as Big Data, AWS, GCP, Azure, GenAI, Data Engineering, Real-Time Data Processing, Cloud-Native & Serverless Architectures. * I have extensive expertise in Scalable System Design (HLD & LLD) with Design Patterns, including CAP, CDN, Cache, Load Balancer, and advanced concepts like Data Mesh, Data Fabric, Data Governance, Data Security, and Data Privacy. * Proficient in LLM, Prompt Engineering, Vector Databases, Agentic AI, LlamaIndex, LangChain, and Model Tuning (LoRA, QLoRA) with RAG solutions, N8N workflow automation. •Expertise in AWS services such as S3, Glue, EMR, EC2, Redshift, Lambda, RDS, Secrets Manager, Amazon Managed Airflow, and CloudWatch, with experience in building end-to-end frameworks using these services. •Proficient in AWS streaming services: SQS, SNS, EventBridge, and Kinesis for real-time data processing. •AWS CloudFormation expertise: Designed and implemented complex CloudFormation templates to automate infrastructure provisioning, using AWS services like Service Catalog. •Data Governance: Proficient in addressing data governance across security, compliance, access control, auditing, and data management using AWS services such as Lake Formation, AWS Glue Data Catalog, IAM, CloudTrail, and KMS. •Spark experience: Extensive work with Spark DataFrames, Datasets, and RDDs, creating end-to-end frameworks using transformations like withColumn, filter, groupBy, orderBy, join, count, and creating temporary views. •Proficient in Python: Skilled in data analysis, manipulation, and processing using Pandas and NLTK, with experience in building frameworks using OOP concepts and leveraging Langchain. •NLP experience: Expertise in Tokenization, Stop Words, Stemming, Lemmatization, Bag of Words, and Word2Vec. •GenAI and Vector DB knowledge: Experience with OpenAI, Gemini, Hugging Face, Groq Cloud, and Vector DBs such as FAISS, ChromaDB, Pinecone, and LlamaIndex. prompt engineering with few-shot, one-shot, and hybrid prompts. •Deep Learning basics: Familiarity with ANN, RNN, LSTM, Transformers, and BERT. •Cassandra, Snowflake and DynamoDB experience: Proficient in Cassandra Data Replication (Replication Factor, Replication Strategy) and DynamoDB (on-demand and provisioned modes). Skilled in managing Snowflake External Tables and automating data ingestion using Snowpipe. •Experience in building High level design and low level design.

Experience

Barclays

Lead Data Engineer

Jul 2021Present · 4 yrs 8 mos · Pune, Maharashtra, India

  • Creating Data Pipeline, Data Mart and Data Recon Fremework for Anti Money Laundering Financial Crime Data
  • Working on Financial Crime / Fraud Detection Data.
  • Develop and Automate end to end Data pipeline using Big Data Technology and cloud AWS.
  • Working on Barclays cards data platform creating framework for card platform Customer risk rating.
  • Implemented data quality check for data validation.
  • Created reconciliation framework for Barclays cards platform customer risk rating.
  • Work on new customer integration and data
  • Worked on different LOB like Business card, Loan, Consumer card, Deposit Card.
  • Worked on patners Integratation like Gap data.
  • Working on cloud data migration project technology use like Direct connect,Data sync DMS etc.
  • Creating datalake for a single source of truth (SSOT) is the practice of aggregating the data from many systems within an organization to a single location.
  • Created different framework like Data Pipeline, Data Quality,Data Reconciliation, Data Archival, Data Governance( using pyspark and AWS )
  • Exposure on AWS cross account and multicloud.
Big DataAWSData PipelineData GovernanceData Quality

Connectwise

Senior Data Engineer

Jul 2020Jul 2021 · 1 yr · Pune, Maharashtra, India

  • Working as a senior software Developer in Big Data.
  • Technology Stack: AWS (S3, EMR, Ec2, Lambda, Kinesis, Cloudwatch, Managed Airflow, SecretManager, Redshift etc), AZURE( Azure Sql), Snowflake, Cassandra, Spark, Python, Boto3.
  • Create Data Quality Framework using Pyspark, S3, Snowflake, EMR, EC2, SecretManager, Airflow, Cloudwatch, Cassandra
  • Created Database Migration ( snowflake ) Framework using Pyspark Sequence creation, DDL Generation, Time travel, Data Copy and many more
  • Created Reconciliation Framework For Database Migration and Data Processing it has count, row by row and Object row by row comparison Feature
  • Managed Data Governors and Data Profiling
  • Poc on Aws Managed Airflow, AWS Glue, Aws Lambda
  • Poc on Image or face Detestation using dlib, OpenCV, and Python
  • Poc on Word Cloud using Python, Matplotlib etc
  • Poc on Snowflake Share inbound and outbound, Create External table using AWS and Azure, Role and many more
  • Poc on statistical analysis data analysis using python and Matplotlib
AWSPysparkSnowflakeCassandraData Quality FrameworkBig Data

Figmd, inc.

Data Engineer

Feb 2018Jul 2020 · 2 yrs 5 mos · Pune, Maharashtra, India

  • > Transforming United States Healthcare data in Scala and providing to CMS (Centers for Medicare and Medicaid Services, USA) on AWS environment. Handling data migration to transform data using Spark RDDs, Dataframes & dataset, Cassandra and Hive.
  • >Working on Differient Data Pipeline Processing like Validate,Filter,Lookup,Transform.
  • >Extensively Writing scripts to load data to Postgres, Hive & Cassandra which are our migrated databases.
  • > Working on the module, writing Spark Jobs in IntelliJ IDE to pull Electronic Health Record from multiple sources like MS SQL Server,Hive,Cassandra and process using Spark and scala.
  • > Handled the module for data - backup in s3 bucket thereby creating a Data lake in Hive, storing tables in parquet & ORC format with compression codec as SNAPPY for the data of several registries under one umbrella. Also dealt with the policies to take the unimportant data to Amazon Glacier.
  • >Handling data migration and transform using spark,rdd, dataframe, dataset using scala and python.
ScalaSparkAWSCassandraData PipelineBig Data

Ibm

Data Engineer

Aug 2016Feb 2018 · 1 yr 6 mos · Pune Area, India

  • Daily we getting near about 15 to 20 GB of data from upstream dump that data in underline database.
  • Efficiently manage this data we implementing Partition, Indexing, performance tuning and purging on database.
  • Dump Data in Hdfs and uncover the information using data processing tool like Pig, Hive and Spark.
  • Descriptive and Predictive analysis of Data and shared required analysis with Vodafone R&D, Marketing, Security & TRAI Team.
  • VAS Services Configuration database configured according to Client requirement(Dump and process semi structure data using pig)
  • Implement Security feature in application and configured in database for e.g. Captcha, Ussd, Dynamic Captcha ,OTP etc.
  • Working with procedures, functions, packages and Exceptions in PL/SQL.
  • Requirement gathering, Coding, Integration.
  • Development of code and query as per Demand
  • Tracking and Closure of all SR/RFC/Tickets.
  • Root cause analysis for Issue/defect in production end.
  • Working with product Deployments and Deliverable.
PL/SQLHadoopSparkData ProcessingData Engineering

Tech mahindra

Software Developer

Sep 2014Apr 2016 · 1 yr 7 mos · Pune Area, India

  • Working with procedures, functions, packages and Exceptions in PL/SQL.
  • Requirement gathering, Coding, Unit testing, Integration.
  • Incident handling.
  • Respond to High Severity issues within stipulated timelines and within agreed SLAs.
  • Problem Management.
  • Root cause analysis for Issue/defect in production end.
  • Working with Change and Request Management
  • Working with product Deployments and Deliverable.
  • Co-coordinating with multiple teams and following them up till the closure of issues
PL/SQLData ProcessingIntegrationSoftware Development

Education

Sinhgad school of Computer studies

Master’s Degree — Computer Engineering

Jan 2010Jan 2013

Mangalvedhakar institute of management

Bachelor’s Degree — Computer Science

Jan 2008Jan 2010

Stackforce found 100+ more professionals with Big Data & Aws

Explore similar profiles based on matching skills and experience