V

Vino Duraisamy

Product Manager

San Francisco, California, United States10 yrs 9 mos experience
Most Likely To SwitchAI Enabled

Key Highlights

  • Expert in Data Engineering and AI workloads.
  • Developed automated training pipelines for ML models.
  • Created scalable data pipelines for enterprise analytics.
Stackforce AI infers this person is a Data Engineering and AI specialist with a focus on enterprise solutions.

Contact

Skills

Other Skills

Apache OozieBig DataBig Data AnalyticsBusiness Intelligence (BI)Churn AnalysisCritical ThinkingCustomer AnalyticsData MiningData ScienceData WarehousingDatabricks Unified Data Analytics PlatformDevOpsHigh Performance OrganizationsKerasMatplotlib

About

πŸ“Œ "In God we trust; all others must bring data." Data & AI practitioner with a unique blend of data engineering, machine learning, artificial intelligence and business analysis expertise. Currently, as a Developer advocate at Snowflake I work on Data Engineering and AI workloads. In my professional experience, I have worked on end-to-end data & AI projects that involved Data Engineering, Data Modeling, Machine Learning Model Deployment, Data Visualization, and Analytics Framework Development for solving business problems. Most recently, as a Data Engineer at Applied ML Search team at Apple, I built a search metrics data store for retail/e-commerce data, and developed automated training pipelines for search query intent classification model and learn-to-rank ML models in production. Previously, as an AI Engineer at IproTech, I worked on a custom NLP model for Named Entity Recognition that identifies and redacts personally identifiable and sensitive information from unstructured text documents. Prior to that, I worked as a Data Engineer at Nike creating robust and scalable enterprise data pipelines for the Consumer Data Analytics & Engineering (CDEA) team. βœ… Programming: Python (Numpy, Pandas, NLTK, Spacy, Scikit-Learn), PySpark, SQL. βœ… Big Data Tools & Frameworks: Apache Spark, Snowflake, Hadoop, ETL data pipelines, Apache Airflow, Hive, AWS (EMR, S3, EC2, Lambda, SNS, Glue, Redshift), lakeFS βœ… VCS, DevOps & Misc Tools: PyCharm, VS Code, Git, Jira, Jenkins, Docker βœ… Statistics: Inferential Statistics, Experimental Design, Hypothesis Testing (A/B Testing), Regression Analysis βœ… Machine Learning: Regression Modeling, Random Forest, XGBoost, kNN Classifier, K-means Clustering, Feature Extraction (PCA, Factor Analysis), Natural Language Processing (Text Analytics – PII, PHI Extraction), Convolutional Neural Network. βœ… Business Domain Expertise: Enterprise Data Analytics (Data Warehousing), Sales and Marketing Analytics, Customer Segmentation, Customer Lifetime Value and Retention Analysis, Customer Success KPIs, Product Analytics Reach me at: βœ‰οΈ vinodhini.sd@asu.edu

Experience

Snowflake

2 roles

Senior Developer Advocate - Data & AI

Promoted

Aug 2025 – Present Β· 7 mos

Developer Advocate - Data & AI

Jul 2023 – Aug 2025 Β· 2 yrs 1 mo

Towards data science

Data Engineering Writer

Jan 2023 – Present Β· 3 yrs 2 mos Β· Remote

Treeverse - the lakefs creators

Developer Advocate - Data Engineering, Machine Learning, MLOps

May 2022 – Jun 2023 Β· 1 yr 1 mo Β· San Francisco Bay Area

  • lakeFS is an open-source tool that offers data versioning at scale for object stores - think "git for data", only it scales for 100s of petabytes datalakes.
  • ● Start managing data the way you manage your code.
  • ● lakeFS sits on top of your object store and provides git-like capabilities such as commit, branch, revert or merge all via UI, CLI or API.
  • ● The ecosystem of tools you have today can access the data via lakeFS the same way they access it today. Easy peasy! Yep, I said it.

Apple

Data Engineer - Applied AI & Search

Apr 2021 – May 2022 Β· 1 yr 1 mo Β· Sunnyvale, California, United States

Ipro

AI Engineer - Language Models (NLP)

Sep 2020 – Mar 2021 Β· 6 mos Β· Tempe, Arizona, United States

  • Tools & Languages Used: Python (Spacy), Named Entity Recognition Models (Decision Tree, Random Forest, CNN), Docker, Jenkins
  • ● Developed and trained a Spacy based Named Entity Recognition model to identify and redact personally identifiable information from unstructured text data.
  • ● Preprocessed and cleaned a corpus of 1M+ documents (emails, tweets, wikipedia and news articles) using Python.
  • ● Engineered features for different document types and built an integrated feature pipeline for model training.
  • ● Achieve an average F1-score of 80% for PII data and deployed the model as a python package in production.
  • ● Worked with product managers and development teams to ensure accurate integration of the model into the product.

C1x inc.

Data Engineer (client: Nike)

Aug 2020 – Oct 2020 Β· 2 mos Β· San Jose, California, United States

  • Tools & Languages Used: Python, Spark, Hive, AWS (S3, EMR, DynamoDB), Airflow, Hadoop, Docker, Jenkins
  • ● Developed robust, scalable data pipelines for data privacy compliance module in Airflow for Customer Data Engineering & Analytics team at Nike.
  • ● Created several DAGs to ensure hourly ingestion of customer data from AWS S3 buckets into Hive tables.
  • ● Designed and built re-usable libraries in PySpark to ensure data quality and to support data engineering, and downstream analytics workflows.
  • ● Continuously monitored and improved data pipelines by analyzing bottlenecks, and implemented efficient solutions.
  • ● Refactoring & optimizing the code for improved reliability, performance, simplicity and maintenance.

Medium

Technical Writer - Data Engineering, AI, ML, Statistics

May 2020 – Present Β· 5 yrs 10 mos Β· Remote

  • ● Publishing articles on a variety of topics including Data Engineering, AI, ML, Statistics and Data Analytics.

Ipro

AI Engineer - Language Models (NLP)

Nov 2019 – May 2020 Β· 6 mos Β· Phoenix, Arizona Area

  • Tools used: Python (SpaCy, Gensim, Regex), Prodigy, AWS S3 & EC2
  • ● Developed a prototype PII (Personally Identifiable Information) extraction and Named Entity Recognition model for Email corpus using Python.
  • ● Leveraged SpaCy, Gensim and regex libraries for data cleaning and pre-processing to improve data quality.
  • ● Built SpaCy based NER model and rule based PII extraction engine to identify, tag and redact PII data including health, finance, security and other sensitive information (such as Social Security Number, Bank Account Number, Credit Card number, etc. )
  • ● Re-trained SpaCy English language model β€˜en_core_web_lg’ on Email corpus using the data annotation and training tool Prodigy for improved model performance.

Arizona state university

Explainable AI Researcher

Sep 2019 – May 2020 Β· 8 mos Β· Tempe, Arizona, United States

  • ● Worked with Dr. Asim Roy from Dept of Information Systems on improving explainability and interpretability of deep learning models.
  • ● Built a 4 layer Convolutional Neural Network for MNIST handwritten digit recognition with 99.9% accuracy as the base model for analysis.
  • ● Analyzed the filters, pooling layers and inter-connected layers of convolutional neural networks and visualized the activation values at each layer to understand underlying abstractions at each level.
  • ● Leveraged Activation maximization to generate the input image that maximizes activations of particular neuron or a group of neurons, thus revealing the abstractions captured by the neurons.
  • ● Used Saliency maps to understand the part of input image that contributes to the activations.

Adrenalin esystems ltd

Analytics/ Data Engineer - Sales & Marketing

Oct 2018 – Jun 2019 Β· 8 mos Β· Chennai Area, India

  • Tools used: Python, Tableau, MySQL, Excel
  • ● Gathered data on customers, competitive products and market place, and consolidate information into actionable items, reports and presentations.
  • ● Built a segmentation model to classify prospective customers using different factors enabling sales teams to derive efficient strategies for different prospect groups, thus decreasing the average sales cycle.
  • ● Developed a customer churn prediction model to identify the top 3 factors contributing to customer churn, enabling the account managers to work on retention strategy. This increased the retention rate and improved the overall customer satisfaction score as well.
  • ● Created Tableau dashboards and periodic reports for senior leadership to track customer churn metrics.

Datatracks

Analytics/ Data Engineer - Partnerships and Alliances

May 2018 – Aug 2018 Β· 3 mos Β· Chennai Area, India

  • Tools used: Python, Tableau, Excel Stat Tools
  • ● Conducted detailed market research to explore potential partnership opportunities in European countries and presented findings to top leadership. Leveraged Tableau and Pareto charts for data analysis and reporting.
  • ● Prototyped a channel partner recommendation model using existing partner persona to identify right partnership opportunities in the European region thereby increasing the conversion rate. Honored with Impact Award.
  • ● Performed qualitative and quantitative analysis of user experience on the website to understand customer behavior and pain-points for effective customer engagement.
  • ● Leveraged Web Analytics UX tools like Session replay, Heat Maps and Conversion Analytics through metrics like bounce rate, time on site and time on page to study the user journey on the website.

Netapp

3 roles

Software Engineer II - Customer Success Ops

Apr 2017 – May 2018 Β· 1 yr 1 mo

  • Tools & Technologies: Python, C++, MySQL, Cassandra, Tableau
  • ● Resolved customer issues in production environment for SMO (Snap Manager for Oracle), SDU (Snap Drive For Unix), WFA (Workflow Automation) and OCI (On Command Insight) applications and released priority patches for customers with mission critical applications.
  • ● Worked closely with customers and end users to understand and document the business use cases, collaborated with Product Management and Development teams to include Customer RFE (Request for Feature Enhancements) in the upcoming product releases.
  • ● Recognize patterns in product issues faced by customers through deep-dive analysis of support ticket data across support centers and offer insights to product managers thus improving product stability.
  • ● Identified critical accounts based on the number and severity of customer cases, to prioritize critical accounts resulting in improved support experience.
  • ● Collaborated with Sales Account Managers to evaluate customer’s data center environment and offer technical recommendations to upgrade the product stack, as required.
  • ● Analyzed business practices ensuring best practices were captured and implemented consistently.
  • ● Hands-on experience in building dashboards for Customer Success KPIs and Support Center Score cards for various stakeholders.

Software Engineer I - Customer Success Ops

Aug 2015 – Apr 2017 Β· 1 yr 8 mos

  • Tools & Technologies used: Python, C++, MySQL, Cassandra, Tableau
  • ● Worked on WFA (On Command Workflow Automation) software that automates storage provisioning, configuration, management and other service tasks in a NetApp Data Center.
  • ● Analyzed and segmented customer issues by products and modules, presented actionable insights to the product management team enabling them to improve certain product workflows.
  • ● Built and interactive dashboard and generated ad hoc reports of various customer support performance metrics using Tableau and Excel for top leadership.
  • ● Single handedly fixed customer issues in WFA production environment and released product patches within stipulated SLA (Service Level Agreement) for several Fortune 500 customers across the globe.
  • ● Extracted data logs from multiple data sources, performed troubleshooting, Root Cause Analysis and documentation of product issues in Support Center Knowledge base enabling NetApp customer community.
  • ● Trained L1 and L2 support teams through Product training and created effective product documentations to help improve customer support KPIs

Software Engineer Intern

Dec 2014 – Jul 2015 Β· 7 mos

  • ● Understood NetApp's proprietary file system WAFL (Write Anywhere File Layout) and the detailed working of data manageability operations like Deduplication, Clone, Compression, Consistency Point, Encryption, etc.
  • ● Developed a Perl based tool which runs along with integration tests to monitor the WAFL operations counters (Deduplication, Clone, Compression, Consistency Point) in real-time, generated automated reports at the end of tests, offered insights and feedback to improve test case development.
  • Tools used: Perl

Young leaders for active citizenship (ylac)

Research Analyst - Public Policy

Feb 2017 – Mar 2017 Β· 1 mo Β· Bangalore

  • ● Understood Indian legislation, framework of public policies and the role of Governmental and Non-governmental organizations, Corporations and other stakeholders involved in policy making in India.
  • ● Performed extensive research and analysis of the existing Rainwater harvesting policy in Karnataka for the office of the Member of Parliament (Upper House) Dr. Rajeev Gowda.
  • ● Presented a revised policy framework mandating rainwater harvesting in major cities of Karnataka and developed a detailed advocacy plan to increase the voluntary adoption of rainwater harvesting methods in the state.

Indian institute of technology, bombay

Open Source Contributor, Scilab Textbook Companion Project

Dec 2015 – Jan 2016 Β· 1 mo Β· Mumbai

  • ● Programmed all the examples of the book β€œSignals and Systems” written by Dr.V.Krishnaveni and Dr.A.Rajeswari.
  • ● Visualized all the answers using Scilab, a numerical computational package and an open-source alternative of Matlab.
  • ● The code was extensively reviewed and added to Scilab documentation.
  • ● The internship was funded by FOSSEE (Free and Open Source Software for Education) project, IIT Bombay. Scilab Textbook Companion Project: https://scilab.in/Textbook_Companion_Project

Caterpillar inc.

Business Intelligence Analyst

May 2014 – Jun 2014 Β· 1 mo Β· Chennai Area, India

  • Summer internship opportunity at Caterpillar India Pvt ltd was offered to the winning team of FIRST Tech Robotics Challenge, Chennai Chapter.
  • ● Understood Caterpillar's product portfolio and In-vehicle networking protocols like CAN, Flex-Ray, Ethernet, LIN, etc.
  • ● Built interactive charts and dashboard using Tableau to capture the Quality Assurance Test metrics for Engineering Managers.

Education

W. P. Carey School of Business – Arizona State University

Master of Science - MS β€” Business Analytics

PSG College of Technology

B.E. β€” Electronics and Communications Engineering

Stackforce found 100+ more professionals with Apache Oozie & Big Data

Explore similar profiles based on matching skills and experience