Varun Prashant Gangal — AI Researcher
At Patronus AI, I do research on language generation & LLM evaluation, with a specific focus on devising novel envs to further abilities of Large Language Models (LLMs) as well as more generally, LLM-driven agentic AI frameworks and systems. Before Patronus AI, I was an AI/NLP/LLM researcher at Amazon AGI , NYC researching LLM evaluation & post-training [reasoning benchmarks, reward models, LLM Judges, long context ability]. I was a contributor to the Nova family of LLMs (https://tinyurl.com/5ea726cv) Before Amazon AGI, I was a Research Scientist at ASAPP Inc, NYC from Dec'22 - Aug'24, performing both research & product dev in the broad areas of generative AI/ NLP [Natural Language Processing], exploring & solving problems in LLMs, efficient training & finetuning, prompt optimization, data augmentation inter alia. Formerly, I did my PhD from 2016-22 at the CMU Language Technologies Institute (LTI) , advised by Prof Ed Hovy, with research foci in Natural Language Generation and Data Augmentation. Some of my recent and current work has been focussed on: 1. How well LLMs can meta-critique LLM-driven agentic traces (https://huggingface.co/papers/2505.08638) 2. Aligning LLMs for Task Dialog [Multi Turns + Tool-Call] Settings, to appear at Findings of ACL 2025 (Preprint here: https://arxiv.org/abs/2409.04617) 3. Efficient ML / Accelerating NN Training E.g., DYAD, a blocksparse approx. for neural net linear layers, NEURIPS'23 WANT WS (Paper here: https://openreview.net/forum?id=obE6BSiUjt OR surf X thread at https://x.com/VarunGangal/status/1727366831575347468) 4. Creative NLG/ Generative AI Generating creative language artifacts such as tongue twisters (accepted at EACL 2023) (Paper: https://arxiv.org/pdf/2209.06275.pdf) & personification (accepted at COLING'22) (Paper: https://aclanthology.org/2022.coling-1.547/ ) 5. LLM Robustness, Alignment & Safety a) Detecting hallucinations through counterfactual data synthesis (Our paper at ACL 2024 Findings: https://openreview.net/forum?id=T1kZ0tdOtZM [Anon preprint]) b) Designing jailbreak filters, unit testing frameworks inter alia to ensure reliable, safe and compliant behaviour of LLM powered AI systems c). Detecting Euphemisms - our team's system EUREKA placed first in the Figlang '22 task, co- located with EMNLP'22 (Paper: https://aclanthology.org/2022.flp-1.15.pdf) You can access a list of [and links to] my research on my Google Scholar page (https://scholar.google.com/citations?user=rWZq2nQAAAAJ&hl=en)
Stackforce AI infers this person is a leading expert in AI/NLP with a focus on generative models and LLMs.
Location: Jersey City, New Jersey, United States
Experience: 12 yrs 11 mos
Career Highlights
- Expert in LLM evaluation and generative AI frameworks.
- Contributed to multiple high-impact AI research publications.
- Strong background in data augmentation and NLP techniques.
Work Experience
Patronus AI
Research Scientist (10 mos)
Amazon AGI
Applied Scientist II (8 mos)
ASAPP
Research Scientist (1 yr 8 mos)
Ai2
Summer Research Intern (3 mos)
Summer Intern (3 mos)
Snap Inc.
Summer Research Intern (3 mos)
Carnegie Mellon University - School of Computer Science - Language Technologies Institute
Teaching Assistant, Neural Networks for NLP (4 mos)
Teaching Assistant, Grammars & Lexicons (4 mos)
PhD [Defended 30th September] (6 yrs)
Indian Institute of Technology, Madras
Teaching Assistant, Introduction to Machine Learning MOOC, NPTEL (5 mos)
Teaching Assistant, Reinforcement Learning (5 mos)
Teaching Assistant, Introduction To Machine Learning (4 mos)
IBM Research India
Research Intern (3 mos)
Microsoft
Summer Intern at Bing Ads (2 mos)
TCS Innovation Labs
Summer Intern (2 mos)
The Fifth Estate, IIT Madras
Correspondent, Science Writer (3 yrs)
Education
Doctor of Philosophy (Ph.D.) at Carnegie Mellon University
B.Tech + M.Tech (Dual Degree) at Indian Institute of Technology, Madras