Zixuan Song

Associate Consultant

San Diego, California, United States0 mo experience
AI EnabledAI ML Practitioner

Key Highlights

  • Proven expertise in AI agents and large language models.
  • Developed innovative frameworks for multimodal data processing.
  • Achieved significant performance improvements in AI-driven projects.
Stackforce AI infers this person is a Fintech-focused AI researcher with strong capabilities in multimodal AI and data analysis.

Contact

Skills

Core Skills

Ai AgentsLarge Language Models (llm)Retrieval-augmented Generation (rag)Artificial Intelligence (ai)Data AnalysisRobotics

Other Skills

Natural Language Processing (NLP)LangChainData MiningDatabaseBusiness AnalysisComputer AnimationC++Computer Vision

Experience

Shanghai ai laboratory

Research Intern

Dec 2025Present · 3 mos · Shanghai, China · On-site

  • DeepResearch Agents
AI Agents

Enuit llc

AI Engineer

Jun 2025Nov 2025 · 5 mos · Houston, Texas, United States · On-site

  • Design and deploy an MCP-based AI agent integrated with Entrade® to automate trade capture, market data ingestion, and real-time portfolio risk analytics; build LangChain workflows with tool orchestration, session memory, structured JSON outputs, and guarded retries for forward exposure, hedge optimization, and P&L reports.
  • Build a production RAG pipeline for desk workflows using FAISS indices on desk docs and market reports; engineer few-shot prompt templates for risk management to improve retrieval precision and model performance.
  • Deploy and fine-tune LLaMA for other AI features, including trade scheduling line extraction and parcel match.
Large Language Models (LLM)AI AgentsNatural Language Processing (NLP)LangChain

Uc san diego

Research Intern

Jan 2025Jul 2025 · 6 mos · San Diego, California, United States · On-site

  • 🧠 NeurIPS 2025 — Towards General Continuous Memory for Vision-Language Models
  • Proposed CoMEM, a plug-and-play continuous memory framework for Vision-Language Models (VLMs), that encodes multimodal and multilingual knowledge into 8-token compressed embeddings for efficient reasoning.
  • Trained a continuous memory encoder by LoRA fine-tuning a Qwen2.5-VL and a shared-parameter Q-Former, updating only 1.2% of weights on a 15.6K self-synthesized VQA dataset, achieving training efficiency on single H100.
  • Surpassed state-of-the-art RAG methods by 7%+ on six knowledge-intensive VQA benchmarks (OK-VQA, A-OKVQA, InfoSeek, etc.) and demonstrated robust cross-lingual generalization in low-resource settings.
  • Enabled cross-modal memory transfer to non-vision LLMs, improving reasoning without altering their architecture.
Large Language Models (LLM)Retrieval-Augmented Generation (RAG)Natural Language Processing (NLP)

Tsinghua university

Undergraduate Student Researcher

Feb 2024Jun 2024 · 4 mos · Beijing, China · On-site

  • Developed a multimodal time-series prediction framework combining financial news headlines and historical prices via time-heterogeneous GNNs and LLMs to forecast ranked daily stock returns on the FNSPID Dataset.
  • Constructed dynamic inter-company relation graphs by integrating 20-day historical price correlations with LLM-extracted sentiment scores from 500 daily news headlines, and optimized a weighted regression–rank loss function to improve both return estimation and ranking accuracy for long/short portfolio construction.
  • Delivered a consistent +4% gain in ACC/ARR over baselines, achieving a Sharpe Ratio of 1.34 on FNSPID backtest.
Data MiningLarge Language Models (LLM)Artificial Intelligence (AI)

China merchants securities co., ltd.

Macro Research Intern

Feb 2024Apr 2024 · 2 mos · Beijing, China

  • Macroeconomic analysis, policy evaluation.
Data AnalysisDatabaseBusiness Analysis

National university of singapore

Research Internship

May 2023Aug 2023 · 3 mos · Singapore, Singapore · On-site

  • Developed a scalable rendering and web display pipeline for the Universal Differentiable Simulator, which enabled efficient visualization of complex simulations for rigid bodies, fluids, and cloth dynamics.
  • Streamlined workflow by integrating Alembic to cache simulation results and incorporating rendering logic, optimizing storage footprint and retrieval speed by 30%.
RoboticsComputer AnimationC++

Education

UC San Diego

Master of Science - MS — Data Science

Sep 2024Dec 2026

Tsinghua University

Bachelor's degree — Software Engineering

Sep 2020Jun 2024

The University of British Columbia

Exchange — Computer Science

Jan 2023Apr 2023

Tsinghua University High School

High School Diploma

Sep 2017Jun 2020

Stackforce found 100+ more professionals with Ai Agents & Large Language Models (llm)

Explore similar profiles based on matching skills and experience