Tanmay Gangwani

AI Researcher

Seattle, Washington, United States7 yrs 2 mos experience
Most Likely To SwitchHighly Stable

Key Highlights

  • Expert in large-scale product ranking models.
  • PhD specialization in deep reinforcement learning.
  • Experience with post-training large language models.
Stackforce AI infers this person is a Senior Applied Scientist in AI/ML for e-commerce.

Contact

Skills

Core Skills

Large Language Models (llm)Information RetrievalDeep LearningReinforcement Learning

Other Skills

Applied Machine LearningComputer ArchitectureHigh Performance ComputingHigh Performance Computing (HPC)Machine Learning AlgorithmsOptimizationPyTorchSearch Engine RankingSoftware DevelopmentSoftware Engineering

About

I am a Senior Applied Scientist at Amazon, where I develop large-scale product ranking models for the e-commerce platform. My work focuses on designing architectures and algorithms that infer customer preferences and deliver personalized search results, with an emphasis on post-training large language models (LLMs) to align with user preferences. Before joining Amazon, I earned my PhD from UIUC, specializing in deep reinforcement learning and imitation learning algorithms. Personal website (not updated since graduation): https://tgangwani.github.io

Experience

Amazon

2 roles

Senior Applied Scientist

Promoted

Dec 2024Present · 1 yr 3 mos · Seattle, Washington, United States · On-site

  • Designing architectures to infer implicit user preferences and scale personalized search
  • Post-training LLMs to better reflect user preferences
  • Leveraging LLMs to improve evaluation strategies for personalized product ranking models
Large Language Models (LLM)Information Retrieval

Applied Scientist II

Oct 2021Nov 2024 · 3 yrs 1 mo · Seattle, Washington, United States · On-site

  • Deep learning models for search product ranking and search query autocomplete
  • Multi-objective optimization and learning-to-rank algorithms
Deep LearningInformation Retrieval

D-wave systems inc.

Research Internship

May 2019Aug 2019 · 3 mos · Vancouver, British Columbia, Canada

  • Model-based Reinforcement-learning for Industrial Process Control -- Worked on model-based RL algorithms for learning control in industrial process applications (e.g. manufacturing plants). We investigated techniques for learning robust deep-learning models of the environment dynamics that are accurate in long-horizon predictions and could be used to learn control policies in the offline-RL setting.

Uber ai

Research Internship

May 2018Aug 2018 · 3 mos · San Francisco Bay Area

  • Imitation Learning in partially observable Markov decision processes (POMDPs) - We investigated deep-learning architectures and algorithms for imitation learning (from expert demonstrations) in POMDP environments using adversarial methods. Please refer to the paper for details - https://arxiv.org/abs/1906.09510

Amd

Research Internship

Jun 2015Sep 2015 · 3 mos · Bellevue, Washington

  • Worked with the GPU modeling team on hardware techniques to make the GPU more amenable to running complex, irregular codes as opposed to only regular, embarrassingly parallel applications. We proposed modifications to the way threads are handled inside a GPU by augmenting the command-processor.
  • Implemented new hardware modules in the AMD GEM5 simulator to evaluate the potential benefits on OpenCL applications under the HSA (Heterogeneous System Architecture) platform.

Intel corporation

Software Engineer

Aug 2011Aug 2013 · 2 yrs · Bangalore, India

  • I had the opportunity to contribute to the first two iterations of the Intel Xeon Phi cards, aimed at the high performance computing (HPC) market. Our team handled the performance analysis and modeling of scientific and numeric codes (like Lattice Boltzmann, Molecular Dynamics, Weather Forecasting) on the Knights Corner silicon, and on the Knights Landing performance simulator. Of particular interest were improving the parallelization efficiency of the codes on the many-core chips using compiler and programming-model techniques, and also alleviating hardware-level scalability bottlenecks such as contention at shared resources (e.g. DRAM buffers).

University of new brunswick

Research Internship

May 2010Jul 2010 · 2 mos · Fredericton, Canada

  • Selected for the MITACS Globalink Research Internship program (https://www.mitacs.ca/en/programs/globalink)
  • Worked with Prof. Lichen Chang in the Sustainable Power Research Group on the design of a MOSFET-based amplifier board to handle the dynamic load from a wind turbine generator prototype.

Education

University of Illinois Urbana-Champaign

Doctor of Philosophy - PhD — Computer Science

Aug 2016Aug 2021

University of Illinois Urbana-Champaign

Master of Science (MS) — Electrical and Computer Engineering

Jan 2013Jan 2016

Indian Institute of Technology, Kanpur

Bachelor's degree — Electrical Engineering

Jan 2007Jan 2011

BBPS New Delhi

High school

Jan 2005Jan 2007

Stackforce found 100+ more professionals with Large Language Models (llm) & Information Retrieval

Explore similar profiles based on matching skills and experience