Vivek Kulkarni

CEO

Bengaluru, Karnataka, India4 yrs 8 mos experience
AI EnabledAI ML Practitioner

Key Highlights

  • Expert in designing data-intensive architectures.
  • Proficient in end-to-end machine learning solutions.
  • Strong background in cloud technologies and data engineering.
Stackforce AI infers this person is a Data Engineering and Machine Learning expert in the SaaS industry.

Contact

Skills

Core Skills

Ai AgentsMicrosoft AzureDatabricksAmazon Web Services (aws)Machine Learning

Other Skills

Azure AI FoundryPython (Programming Language)Lean Six SigmaContinuous Integration and Continuous Delivery (CI/CD)AI LiteracyAI for BusinessBusiness Process ImprovementRoot Cause AnalysisBusiness AnalysisMicrosoft Power BI

About

Data Scientist | Machine Learning Engineer | Data Engineer | Designing Data-Intensive Architectures | Spark, AWS & Azure | Lean Six Sigma | Yellow Belt πŸš€ Data Scientist and Architect with a Master’s in Data Science (USA), specializing in designing data-intensive application architectures that drive business intelligence. My expertise lies in bridging the gap between statistical theory and production-grade engineering. I focus on building end-to-end Machine Learning solutionsβ€”from exploratory data analysis and feature engineering to deploying scalable models in the cloud. Unlike traditional data scientists, I also manage the underlying infrastructure, ensuring that high-performance ETL pipelines and CI/CD workflows support robust ML operations. I love solving data challenges β€” from designing ETL/ELT pipelines to integrating DevSecOps and CI/CD automation β€” turning raw data into business insights and innovation. πŸ’Ό π–π‘πšπ­ 𝐈 𝐃𝐨 𝐁𝐞𝐬𝐭 β€’ Developing predictive models and algorithms to solve complex business problems. β€’ Build & maintain robust ETL/ELT pipelines using Apache Spark, Databricks & Python 🐍 β€’ Design data models, perform schema design & manage versioned migrations 🧱 β€’ Clean, transform & optimize massive datasets for Analytics πŸ“ˆ and Machine Learning πŸ€– β€’ Integrate solutions with AWS ☁️ and Azure 🌐 cloud ecosystems β€’ Implement CI/CD pipelines using GitHub Actions & Infrastructure-as-Code frameworks βš™οΈ β€’ Automate data workflows via REST APIs πŸ”— & event-driven jobs 🧩 🧰 π“πžπœπ‘ 𝐓𝐨𝐨π₯𝐛𝐨𝐱 Python 🐍 | Machine Learning & Statistics πŸ“ | SQL πŸ—ƒοΈ | NoSQL πŸš€ | Apache Spark / PySpark ⚑ | Databricks 🧱 | AWS ☁️ | Azure Data Services 🌐 | GitHub CI/CD πŸ” | GitHub Actions βš™οΈ | REST API πŸ”— πŸ—οΈ πŸ” 𝐊𝐧𝐨𝐰𝐧 𝐟𝐨𝐫 π›πžπ’π§π : βœ… An autodidact with endless curiosity βœ… A problem-solver who loves clean architecture πŸ’‘ βœ… Detail-oriented πŸ‘οΈ, results-driven 🎯 βœ… Focused on designing resilient, production-grade data backbones that power enterprise AI solutions πŸŽ“ π‚π„π‘π“πˆπ…πˆπ‚π€π“πˆπŽππ’ 🧠 AWS Certified Cloud Practitioner πŸ“˜ Microsoft Certified Azure Data Fundamentals (DP-900) πŸ—οΈ Microsoft Certified Azure Data Engineer Associate (DP-203) πŸ”₯ Databricks Certified Data Engineer Associate πŸ›‘οΈ GitHub Foundations βš™οΈ GitHub Actions 🌱 Always eager to take on challenging roles that blend Data Engineering, ML-readiness, and Cloud Innovation.

Experience

Microsoft

Technical Solution Manager

Oct 2025 – Present Β· 5 mos Β· Bengaluru, Karnataka, India Β· Hybrid

  • ● Architecting data-driven solutions for the Cloud Controls and Procurement Execution (CSPE) and CMOF domains, leveraging Python, PySpark, SQL, PowerBI, and Microsoft Azure to optimize Data Centre hardware and software capabilities.
  • ● Developing and deploying autonomous AI Agents using Microsoft Copilot Studio and Azure AI Foundry to automate complex supply chain workflows and enhance decision-making speed.
  • ● Integrating Lean Six Sigma principles and Kaizen methodologies into engineering workflows, resulting in measurable process improvements and reduced operational latency.
  • ● Acting as the Centre of Excellence lead for emerging AI technologies; responsible for rapid prototyping (POC), testing, and deploying scalable AI models into production Azure environments
  • ● Designing robust ETL pipelines to synthesize multi-vertical telemetry data, creating a unified data ecosystem for advanced analytics and reporting.
AI AgentsAzure AI FoundryMicrosoft AzurePython (Programming Language)Lean Six Sigma

Enterprise minds, inc

Senior Data Engineer

Sep 2025 – Oct 2025 Β· 1 mo Β· Pune District, Maharashtra, India Β· Hybrid

Cloudaeon

Senior Data Engineer

Feb 2023 – Jan 2025 Β· 1 yr 11 mos Β· Pune District, Maharashtra, India Β· Hybrid

  • ● Spearheaded the design and development of end-to-end ETL pipelines using Azure Data Factory (ADF) and Databricks, enabling seamless data ingestion, transformation, and integration across diverse enterprise data sources.
  • ● Developed and maintained Python-based backend automation scripts to streamline release management workflows between SharePoint, JIRA, and Azure SQL, reducing manual interventions and improving release turnaround time.
  • ● Extensively utilized Azure Logic Apps to orchestrate event-driven data flows and automate data validation checks, integrated with email notification systems for real-time success and failure alerts.
  • ● Implemented DevSecOps practices by integrating Snyk, TruffleHog, and Fortify into GitHub Actions CI/CD pipelines, ensuring early detection and remediation of code vulnerabilities, and enabling automatic JIRA ticket creation for any issues identified.
  • ● Built and deployed Data Audit and Governance frameworks leveraging GitHub REST APIs and GitHub Actions, automating metadata tracking, repository compliance, and documentation validation processes.
  • ● Collaborated closely with cloud architects, security engineers, and product teams to enhance data quality, access control, and compliance standards across the Azure ecosystem.
  • ● Optimized Databricks workflows by tuning cluster configurations, improving notebook runtime efficiency, and enhancing integration with ADF pipelines for better orchestration and scalability.
  • ● Played a key role in knowledge sharing by documenting best practices and conducting internal workshops on Azure Data Engineering, CI/CD automation, and cloud security integration.
Python (Programming Language)Continuous Integration and Continuous Delivery (CI/CD)DatabricksMicrosoft Azure

Dynamon

Senior Data Engineer

Feb 2022 – Feb 2023 Β· 1 yr Β· Pune District, Maharashtra, India Β· Remote

  • ● Designed and implemented an AWS Lambda–based event notification system to alert engineering and operations teams on Slack upon successful completion of ECS task executions, leveraging Python (boto3) and Slack APIs to streamline real-time communication.
  • ● Developed and maintained containerized applications using Docker, building and deploying optimized images through GitLab CI/CD pipelines, ensuring reliability, security, and minimal downtime across multiple AWS environments.
  • ● Architected a comprehensive infrastructure automation framework using AWS CloudFormation, writing modular JSON and YAML templates to provision and manage EC2, ECS, ECR, Lambda, and IAM resources efficiently.
  • ● Collaborated closely with DevOps and backend engineering teams to integrate CI/CD workflows into microservices architecture, reducing manual deployments and accelerating feature delivery cycles.
  • ● Improved infrastructure scalability and resilience by implementing auto-scaling policies, environment-based parameterization, and IAM least-privilege roles within CloudFormation stacks.
  • ● Set up centralized logging and monitoring using CloudWatch and SNS, enabling proactive issue detection and alerting across critical data pipelines and workloads.
  • ● Contributed to cross-functional discussions on cost optimization, introducing lifecycle rules and on-demand resource scheduling that reduced AWS costs by an estimated 20%.
  • ● Championed documentation and knowledge-sharing initiatives to standardize infrastructure setup practices across the data engineering and DevOps teams.
Python (Programming Language)Amazon Web Services (AWS)Continuous Integration and Continuous Delivery (CI/CD)Databricks

Ces

Lead Data Engineer

Jun 2021 – Feb 2022 Β· 8 mos Β· Chennai, Tamil Nadu, India Β· Remote

  • ● Led the design and implementation of robust ETL solutions for the client Benson Hill Biosystems, leveraging the AWS ecosystem to enable scalable, secure, and cost-efficient data processing workflows.
  • ● Architected end-to-end data pipelines integrating AWS S3, Lambda, Glue, RDS, and Redshift, ensuring seamless data ingestion, transformation, and loading for analytics and reporting use cases.
  • ● Utilized Terraform as an Infrastructure-as-Code (IaC) framework to provision, configure, and manage AWS resources β€” ensuring environment consistency, reusability, and faster deployment cycles.
  • ● Developed, optimized, and orchestrated AWS Lambda and AWS Step Functions to automate ETL workflows, reduce manual intervention, and achieve near real-time data processing.
  • ● Implemented centralized monitoring and alerting using CloudWatch, setting up custom metrics, dashboards, and alarms to proactively detect performance bottlenecks and operational issues.
  • ● Collaborated with cross-functional data, analytics, and DevOps teams to define best practices for data governance, security, and cost optimization within AWS.
  • ● Led peer code reviews, mentored junior engineers, and contributed to internal documentation and reusable Terraform modules to standardize infrastructure deployments across projects.
  • ● Played a key role in migrating legacy ETL scripts to serverless architecture, improving scalability, reducing maintenance overhead, and cutting operational costs by over 30%.
  • ● Delivered well-documented, production-grade data workflows that improved reporting accuracy and supported strategic decision-making for client stakeholders.
Python (Programming Language)Amazon Web Services (AWS)Continuous Integration and Continuous Delivery (CI/CD)Databricks

Rx savings solutions

Senior Data Engineer

Dec 2020 – Apr 2021 Β· 4 mos Β· Overland Park, Kansas, United States

  • ● Expertly executed OCR pipeline operations on documents received from various pharmaceutical and healthcare service provider companies.
  • ● Collaborated closely with the ETL team to ensure smooth and accurate data migration across teams.
  • ● Explored and learnt the latest AWS technologies to provide new capabilities and increase efficiency in the ETL Pipelines.
  • ● Stayed current with the latest AWS technologies and applied them to enhance capabilities and improve efficiency in ETL pipelines.
  • ● Demonstrated expertise in the areas of data management, including integration, modeling, analytics, and reporting by utilizing Python, the AWS ecosystem, and big data tools.
Python (Programming Language)Amazon Web Services (AWS)Continuous Integration and Continuous Delivery (CI/CD)Databricks

United solutions, llc

Senior Data Scientist

Aug 2020 – Dec 2020 Β· 4 mos Β· Rockville, Maryland, United States

  • ● Worked with a Federal Government client to solve large-scale contract document classification and information extraction problems using both supervised and unsupervised NLP algorithms.
  • ● Developed a Flask-based API web application that ingests client documents and accurately classifies them into predefined categories, streamlining document management and review workflows.
  • ● Built multiple independent Proofs of Concept (POCs) for diverse document types to automatically detect and extract key information using PyTesseract and Tesseract OCR in Python.
  • ● Leveraged NLP libraries such as NLTK, spaCy, and FastText to perform text preprocessing, cleaning, stemming, lemmatization, and word encoding for model training and inference.
  • ● Experimented with TF-IDF, Word2Vec, and FastText embeddings to improve text feature representation and classification accuracy.
  • ● Collaborated with data engineers and business analysts to validate model outputs against domain rules and continuously refine accuracy through iterative feedback cycles.
  • ● Delivered insights that helped automate manual contract review processes, reducing processing time and improving document handling efficiency for the client.
Python (Programming Language)Amazon Web Services (AWS)DatabricksMachine Learning

Toyota north america

Data Scientist

Feb 2020 – Apr 2020 Β· 2 mos Β· Dallas-Fort Worth Metroplex

  • ● Worked for marketing team to deliver actionable insights by leveraging historical data and applying various statistical / machine learning algorithms
  • ● Understood business requirements, tested various hypothesis about underlying mechanics of the
  • business processes and wrote clean, efficient code in R and Python.
  • ● Prepared presentations and reports on findings as well as recommended next steps to ensure successful project completion.
Python (Programming Language)DatabricksMicrosoft AzureMachine Learning

Cisco

Data Scientist

Jun 2018 – Jan 2020 Β· 1 yr 7 mos Β· San Francisco Bay Area Β· On-site

  • ● Collaborated with onshore (US) & offshore (Bangalore) analytics teams to design & deliver scalable machine learning solutions addressing Cisco’s enterprise-level quality and reliability challenges, ensuring alignment with global business KPIs.
  • ● Developed and iteratively improved multiple supervised learning models β€” classification (Logistic Regression, Decision Tree, Random Forest, SVM, XGBoost) and regression (Linear, Ridge, Lasso, PLS, Random Forest Regressor) β€” driving measurable accuracy and stability improvements across sprints.
  • ● Performed detailed validation and statistical benchmarking using Accuracy, Cohen’s Kappa, RMSE, ROC-AUC, Precision, Recall, and F1-score to ensure both analytical rigor and business relevance.
  • ● Productionized prototype models built in Python by implementing end-to-end machine learning pipelines on distributed MapR Hadoop clusters β€” integrating automated data ingestion, feature engineering, model training, testing, and deployment for real-time and scheduled batch scoring.
  • ● Created robust data preprocessing workflows including cleansing, missing value handling, outlier detection, class imbalance correction using SMOTE, and feature ranking with permutation importance and SHAP for interpretability.
  • ● Designed model monitoring scripts and performance dashboards using Python, NumPy, and Matplotlib to track drift, stability, and retraining needs.
  • ● Collaborated closely with DevOps teams to integrate ML jobs within enterprise data pipelines using Airflow and Jenkins for scheduling and CI/CD automation.
  • ● Streamlined interoperability between Jupyter Notebooks and MapR HDFS through PyArrow and Hadoop client APIs, enabling efficient reading/writing of large datasets directly from the distributed storage layer.
  • ● Contributed to documentation and internal knowledge sharing sessions focused on best practices for model deployment, versioning, and reproducibility β€” helping accelerate adoption of data-driven decision-making across quality teams.
Python (Programming Language)DatabricksMachine Learning

Education

Mercyhurst University

Master’s Degree β€” Data Science

Jan 2016 – Jun 2018

Dr. Babasaheb Ambedkar Marathwada University, Aurangabad

Bachelor’s Degree β€” Information Technology

Aug 2010 – Mar 2015

M.G.M's Jawaharlal Nehru College of Engineering

Bachelor's degree β€” Information Technology

Aug 2010 – Mar 2015

Stackforce found 100+ more professionals with Ai Agents & Microsoft Azure

Explore similar profiles based on matching skills and experience