Skip to content

yixinhuang48/yixinhuang48

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 

Repository files navigation

Hi, I'm Yixin 👋

Typing SVG

I work on LLM systems, evaluation, and GPU-accelerated ML infrastructure.
RA @ UCSD Hao AI Lab | M.S. Computer Science student
📍 San Diego, CA | Focused on large-scale training, inference, and agent evaluation.

Personal Website Lab Website GitHub Zhihu

Visitor Count

🏆 GitHub Achievements: Pull Shark Quickdraw

"A journey of a thousand miles begins with a single step." — Confucius


💬 Random Dev Joke

Jokes Card


🔬 Research & Systems Interests

  • LLM evaluation & benchmarks (agents, games, scientific reasoning)
  • Large-scale training & inference systems (FSDP, vLLM, Ray, Slurm)
  • GPU efficiency, memory systems, and model parallelism
  • Reinforcement learning for agents (GRPO, NeMo-Gym)

Tech Stack:

Python PyTorch CUDA Docker Ray Slurm Git Linux Jupyter vLLM SGLang NeMo RL Areal


🛠 Selected Projects

🎮 GamingAgent ⭐ 843
LLM/VLM gaming agents and model evaluation through games
→ long-horizon reasoning, memory & perception harnesses
(Doom, Sokoban, Tetris, Pokémon Red)

🔬 VideoScience ⭐ 5
Benchmark for scientific correctness in text-to-video models
→ physics & chemistry concepts, VLM-as-Judge scoring
(CVPR submission)

🤖 NVIDIA NeMo Gym ⭐ 603
Build RL environments for LLM training
→ scalable RL training, reward profiling, GRPO
Integrating Sokoban & Tetris

🌐 lmenv
LLM environment framework for interactive evaluation
→ standardized interfaces for game-based agent testing


GitHub Activity Graph


🧠 Current Focus

  • 🔄 Scaling agent evaluation with interactive environments
  • ⚡ Training & serving efficiency on multi-GPUs
  • 🎯 Reward modeling and RL for LLM agents

📚 Currently Learning

  • Advanced distributed training techniques (FSDP, DeepSpeed)
  • GPU memory optimization and profiling
  • Large-scale RL systems architecture

🔗 Connect with Me

📈 Profile Summary

Profile Summary Card

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published