Profile
| Field | Details |
|---|---|
| Current Position | NVIDIA Director of AI & Distinguished Scientist |
| Team | GEAR Lab Co-Lead (Generalist Embodied Agent Research) |
| PhD | Stanford University (2016-2021) |
| Advisor | Fei-Fei Li |
| Undergraduate | Columbia University (Valedictorian 2016, Illig Medal) |
| Social Media | Active AI communicator (340K+ followers) |
Key Contributions
- GR00T: NVIDIA’s humanoid robot foundation model (Co-Lead)
- Voyager: LLM-based autonomous Minecraft agent
- MineDojo: Minecraft-based AI benchmark (NeurIPS 2022 Outstanding Paper)
- Eureka: Automatic reward function generation with LLMs
- Foundation Agent Vision: Pioneered direction for general-purpose agent research
- OpenAI’s First Intern: 2016 OpenAI Universe/World of Bits project
Research Timeline
Columbia University (2012-2016)
| Year | Work | Impact |
|---|---|---|
| 2016 | Undergraduate Graduation | Valedictorian, Illig Medal |
Internships (2015-2020)
| Year | Organization | Work |
|---|---|---|
| 2015 | Baidu AI Lab | Worked with Andrew Ng, Dario Amodei |
| 2015-2016 | MILA | Research with Yoshua Bengio |
| 2016-2017 | OpenAI | First intern, World of Bits (with Ilya Sutskever, Andrej Karpathy) |
| 2018 | Google Cloud AI | Research Intern |
| 2020 | NVIDIA | Research Intern |
Stanford PhD (2016-2021)
Advised by Fei-Fei Li | Thesis: “Training and Deploying Visual Agents at Scale”
| Year | Work | Impact |
|---|---|---|
| 2018 | Video Understanding | Video comprehension research |
| 2021 | PhD Graduation | Stanford Vision Lab |
NVIDIA (2021-present)
GEAR Lab Co-Lead (Co-founded with Yuke Zhu, February 2024)
| Year | Work | Impact |
|---|---|---|
| 2021.12 | Joined NVIDIA | Senior Research Scientist |
| 2022 | MineDojo | Minecraft AI benchmark, NeurIPS 2022 Outstanding Paper |
| 2023 | Voyager | LLM + Minecraft autonomous exploration |
| 2023 | Eureka | LLM-generated reward functions |
| 2024.02 | GEAR Lab Founded | Co-founded with Yuke Zhu |
| 2024 | GR00T | Humanoid foundation model announced (GTC 2024) |
| 2025.03 | GR00T N1 | Open humanoid VLA (GTC 2025) |
| 2025 | Promotion to Director | Director of AI & Distinguished Scientist |
Major Publications
Foundation Agent
- Voyager (2023) - LLM-based autonomous Minecraft agent
- MineDojo (NeurIPS 2022 Outstanding Paper) - Minecraft AI benchmark
- Eureka (2023) - Automatic reward function generation with LLMs
- VIMA - Multimodal foundation model for robot manipulation
- SURREAL - Distributed reinforcement learning framework
Robotics
- GR00T (GTC 2024) - Humanoid foundation model
- GR00T N1 (GTC 2025) - Open humanoid VLA (2.2B parameters)
Early Career
- World of Bits (ICML 2017) - OpenAI Universe, web browser AI platform
Key Ideas
Voyager (2023)
Core: LLM writes code to autonomously explore Minecraft
Components:
1. Automatic Curriculum - LLM proposes next goals
2. Skill Library - Store/reuse discovered skills
3. Iterative Prompting - LLM fixes code on failure
Results:
- Crafted diamond tools without human intervention
- 3.3x faster skill acquisition than previous methods
Impact:
- Pioneered LLM + game agent research
- Set direction for LLM utilization in Embodied AI
GR00T & GR00T N1 (2024-2025)
Core: General-purpose foundation model for humanoid robots
Architecture (Dual-system):
- System 1: Fast-thinking action model (reflexes/intuition)
- System 2: Slow-thinking VLM (deliberate reasoning)
- Diffusion Transformer for continuous motion generation
GR00T N1 Specs:
- 2.2B parameters (VLM 1.34B)
- 16 action chunks generated in 63.9ms (L40 GPU)
- Trained on real robot data, human videos, synthetic data
- Up to 1024 GPUs used for training (~50,000 H100 GPU hours)
Features:
- Natural language understanding
- Human motion imitation
- Support for various humanoids (1X, Boston Dynamics, Agility, etc.)
Impact:
- World’s first open humanoid foundation model (N1)
- Core of NVIDIA’s robotics ecosystem
- Newton physics engine (collaboration with Google DeepMind, Disney Research)
GEAR Lab Vision
GEAR Lab (Generalist Embodied Agent Research) - Co-founded by Jim Fan and Yuke Zhu in February 2024
Foundation Agent
Goal: One agent performing various tasks in various environments
Games (Minecraft) -> Simulation (Isaac) -> Real Robots
Research Areas:
- LLM-based planning & reasoning
- Vision-language models
- World models trained on internet-scale data
- Robot locomotion & dexterous manipulation
- Large action models (autonomous exploration in games/simulations)
- Simulation & synthetic data pipelines for large-scale learning
Integration with NVIDIA Strategy
- Isaac Sim: Simulation environment
- Omniverse: Synthetic data generation
- Jetson: Edge computing
- GR00T: Foundation model
- Newton: Open-source physics engine
Philosophy & Direction
Research Philosophy
“Agents that succeed in games can succeed in the real world. Generality is the key.”
Research Direction
- 2016-2021: Video understanding (Stanford PhD)
- 2021-2023: MineDojo, LLM + games (Voyager, Eureka)
- 2024-present: Humanoid robotics (GR00T, Physical AGI)
Communication & Influence
Active Social Media Presence
- Twitter/X: 340K+ followers (@DrJimFan)
- AI research commentary, vision sharing
- Industry news curation
- Coined “Physical Turing Test” concept
Public Outreach
- Explains complex AI research accessibly
- Discusses AI research directions
- Model for researcher-public communication
Media Coverage
- Featured in New York Times, Forbes, MIT Technology Review, TechCrunch, WIRED, and more
Awards & Recognition
- NeurIPS 2022 Outstanding Paper (MineDojo)
- Columbia University Valedictorian & Illig Medal (2016)
- Stanford AI Lab alumnus (advised by Fei-Fei Li)
- OpenAI’s First Intern (2016)
- Google Scholar citations: 12,000+
Links
- NVIDIA Profile
- Personal Website
- Twitter/X - 340K+ followers
- Google Scholar