Weixin Liang

Product Engineer

Stanford, California, United States6 yrs 6 mos experience
Most Likely To SwitchAI ML Practitioner

Key Highlights

  • PhD candidate specializing in multi-modal Large Language Models.
  • Introduced innovative Mixture-of-Transformers architecture.
  • Published research on AI documentation practices in Nature Machine Intelligence.
Stackforce AI infers this person is a multi-modal AI researcher with a strong focus on efficiency and ethical AI development.

Contact

Skills

Core Skills

Large Language Models (llm)Multi-modal LlmTrustworthy AiNatural Language Processing (nlp)Multi-modal AiConversational AiData Science

Other Skills

AI EthicsArtificial Intelligence (AI)C (Programming Language)C++Computer ArchitectureComputer VisionDashboardData VisualizationDjangoHTMLJavaLaTeXLinuxMIPS AssemblyMachine Learning

About

On the job market for 2025. Recruiters, please feel free to reach out to discuss opportunities or collaborations! Completing my PhD in Computer Science at Stanford University, specializing in multi-modal Large Language Models (LLMs) and efficient model architectures. Homepage: https://ai.stanford.edu/~wxliang/ My research advances the efficiency frontier of large-scale LLMs, focusing on pretraining optimization and sparse architectures, particularly for Multi-Modal AI. I recently introduced Mixture-of-Transformers (MoT), an innovative architecture that achieves comparable performance to dense models with up to 66% reduction in FLOPs for multi-modal LLM pretraining. This work represents a significant step toward more resource-efficient AI systems. At Meta FAIR, I extended this research through large-scale LLM applications, demonstrating MoT's effectiveness across text, image, and speech modalities. My industry experience spans roles at Amazon, Apple, and Tencent, where I contributed to production-scale ML systems and infrastructure. I am actively exploring opportunities in industry research labs and technology companies to continue advancing multi-modal LLMs and driving innovation in resource-efficient AI. Let’s connect to discuss collaboration opportunities or the future of AI. Stanford PhD CS (2025), Stanford MS EE (2021), Zhejiang University CS (2019).

Experience

Meta

Research Scientist Intern at FAIR, Multi-Modal LLM, Pre-training

Jun 2024Dec 2024 · 6 mos · Menlo Park, California, United States · On-site

  • Keywords: Multi-Modal LLM, LLM Pretraining, Large Language Model, Model Sparsity, MoE (Mixture-of-Experts).
  • How can we reduce pretraining costs for multi-modal models without sacrificing quality? We study this Q in our new work:https://arxiv.org/abs/2411.04996
  • At Meta FAIR, We introduce Mixture-of-Transformers (MoT), a sparse architecture with modality-aware sparsity for every non-embedding transformer parameter (e.g., feed-forward networks, attention matrices, and layer normalization). MoT achieves dense-level performance with up to 66% fewer FLOPs!
  • ✅ Chameleon setting (text + image generation): Our 7B MoT matches dense baseline quality using just 55.8% of the FLOPs.
  • ✅ Extended to speech as a third modality, MoT achieves dense-level speech quality with only 37.2% of the FLOPs.
  • ✅ Transfusion setting (text autoregressive + image diffusion): MoT matches dense model quality using one-third of the FLOPs.
  • ✅ System profiling shows MoT achieves dense-level image quality in 47% and text quality in 75.6% of the wall-clock time**
  • Takeaway: Modality-aware sparsity in MoT offers a scalable path to efficient, multi-modal AI with reduced pretraining costs.
  • Measured on AWS p4de.24xlarge instances with NVIDIA A100 GPUs.
Artificial Intelligence (AI)Large Language Models (LLM)Multi-modal LLMMulti-modal LLM PretrainingMulti-modal LLM Reasoning

Hugging face

Visiting Researcher, Trustworthy AI

Jun 2022Dec 2022 · 6 mos · Palo Alto, California, United States

  • Keywords: Trustworthy AI, NLP (Natural Language Processing), Published in Nature Machine Intelligence (2024)
  • Conducted a large-scale analysis of 32,000+ AI model cards hosted on Hugging Face to evaluate and improve documentation practices in AI.
  • Investigated key aspects of model transparency, including environmental impact, limitations, and evaluation metrics, to promote trustworthy AI development.
  • Employed advanced data science and natural language processing techniques to identify community trends, highlight gaps, and propose actionable improvements to industry documentation standards.
  • Demonstrated the impact of comprehensive documentation on model adoption and usability, bridging the gap between developers and end-users.
  • Collaborated with a multidisciplinary team to publish findings as the first author in Nature Machine Intelligence (2024), establishing a foundational study on AI documentation practices.
  • Advanced expertise in responsible AI, transparency frameworks, and ethical deployment of machine learning models.
Trustworthy AINatural Language Processing (NLP)Data ScienceOpen Source PlatformsAI Ethics

Stanford center for artificial intelligence in medicine and imaging (aimi)

Graduate Research Assistant, Dermatology AI

Jan 2021Jan 2021 · 0 mo · Stanford, California, United States · On-site

  • Keywords: Medical AI, Computer Vision, Healthcare ML, Dermatology, Model Bias, Health Equity
  • Collaborated on development of the Diverse Dermatology Images (DDI) dataset, analyzing 656 pathologically-confirmed dermatological images across Fitzpatrick skin types I-VI (208 FST I-II, 241 FST III-IV, 207 FST V-VI)
  • Evaluated performance of 3 state-of-the-art dermatology AI models (ModelDerm, DeepDerm, HAM10000), documenting ROC-AUC drops from 0.92-0.94 on original test sets to 0.56-0.67 on diverse skin tones
  • Contributed to model fine-tuning experiments that closed the performance gap between FST I-II and FST V-VI, achieving ROC-AUCs of 0.73-0.77 for light skin and 0.76-0.78 for dark skin
  • Analyzed dermatologist consensus labeling accuracy, identifying significant disparities between FST I-II and FST V-VI (sensitivity 0.72 vs 0.59, p < 8.8 × 10⁻⁶)
  • Helped create new benchmark dataset enabling systematic evaluation of AI model performance across skin tones, supporting development of more equitable medical AI systems
  • Co-authored paper published in Science Advances (2022), providing quantitative analysis of bias in dermatology AI models for improving access to care for 3+ billion people globally lacking dermatological services

Stanford office of technology licensing (otl)

Graduate Research Assistant | Foundation Model, AI, NLP

Jan 2021Jan 2021 · 0 mo · Stanford, California, United States · On-site

  • Keywords: Technology Transfer, Data Science, Innovation Analytics, Natural Language Processing
  • Led comprehensive computational analysis of Stanford's technology transfer ecosystem, analyzing 4,512 inventions from 6,557 inventors over a 50-year period (1970-2020) to characterize commercialization patterns and success factors
  • Developed ML classifiers achieving 0.76 AUROC for predicting commercial success from invention abstracts, identifying key linguistic features correlated with revenue outcomes
  • Quantified significant temporal trends including tripling of female inventors (6.5% to 19.7%) and expansion of average team size (2.47 to 3.29 members in biology) from 1995-2020
  • Demonstrated that self-licensed inventions through inventor startups generated highest returns - 100% of inventions with >$10M net income were self-licensed vs only 16% for <$10K income inventions
  • Identified revenue performance disparities across fields, with top categories shifting from electronics pre-2000 to biology/chemistry post-2000, providing strategic insights for technology transfer offices
  • Published findings in Cell Press Patterns (2022), offering first large-scale computational framework for analyzing academic innovation and commercialization patterns

Amazon

Applied Scientist Intern | Multi-Modal AI

Jun 2020Sep 2020 · 3 mos · Sunnyvale, California, United States

  • Keywords: Multi-Modal AI, Multi-Modal Transformer, Neural-Symbolic AI
  • Developed LRTA (Look, Read, Think, Answer), a novel neural-symbolic reasoning framework for multimodal AI that enhances large language models' ability to reason across text and visual inputs. Advanced state-of-the-art performance by 15% in comprehensive reasoning tasks.
  • Key contributions:
  • Architected an end-to-end trainable foundation model integrating scene understanding, semantic parsing, and neural execution modules
  • Pioneered transparent AI reasoning capabilities by implementing human-readable explanations at each reasoning step
  • Created evaluation frameworks to assess and validate model robustness against linguistic perturbations
  • Collaborated with Amazon Alexa AI research team on advancing explainable multimodal foundation models
  • Research published at NeurIPS 2020 (Knowledge Representation and Reasoning Workshop)
  • Amazon Science Page:
  • https://www.amazon.science/publications/lrta-a-transparent-neural-symbolic-reasoning-framework-with-modular-supervision-for-visual-question-answering

Stanford university department of computer science

Graduate Research Assistant

Sep 2019Present · 6 yrs 6 mos · On-site

  • CS PhD@Stanford | Multi-Modal LLM, Pretraining & MoE Research

Tencent

Machine Learning Engineer Intern

Jul 2019Sep 2019 · 2 mos · Beijing, China

Multi-modal AINatural Language Processing (NLP)Transformer Models

Apple

Software Engineer Intern

Apr 2019Jun 2019 · 2 mos · Shanghai City, China

Conversational AINatural Language Processing (NLP)

University of illinois urbana-champaign

Summer Internship

Jan 2018Jan 2018 · 0 mo · Urbana-Champaign, Illinois Area · On-site

  • Keywords: Computer Architecture, Multi-Modal AI, Deep Learning Accelerator (DLA), Memory and Storage Systems.
  • Contributed to DeepStore, a pioneering in-storage neural acceleration system achieving up to 17.7× faster query performance and 78.6× better energy efficiency compared to GPU+SSD systems
  • Collaborated on design of novel channel-level neural accelerator architecture optimized for intelligent query workloads under strict SSD power/resource constraints
  • Helped develop similarity-based query cache system leveraging neural networks, delivering up to 25.9× performance improvement for intelligent queries
  • Participated in comprehensive analysis of storage I/O bottlenecks across visual search, audio matching, and text retrieval applications
  • Co-authored paper published at IEEE/ACM International Symposium on Microarchitecture (MICRO '52), a top-tier computer architecture conference
Data ScienceData VisualizationDashboard

Education

Stanford University

Doctor of Philosophy - PhD — Computer Science

Sep 2021Mar 2025

Stanford University

Master of Science (M.S.) — Electrical Engineering (AI and Software Systems Track)

Jan 2019Jan 2021

Zhejiang University

Bachelor of Engineering - BE — Computer Science

Jan 2015Jan 2019

Zhejiang University

Honnors degree

Jan 2015Jan 2019

University of Illinois Urbana-Champaign

Summer Internship

Jan 2018Jan 2018

Guangdong Experimental High School

High School Diploma — Science & Technology

Jan 2012Jan 2015

Stackforce found 100+ more professionals with Large Language Models (llm) & Multi-modal Llm

Explore similar profiles based on matching skills and experience