Pi Series (Physical Intelligence)

Physical Intelligence's Vision-Language-Action Model Series - Pi0, Pi0.5, Pi*0.6

Author’s Note

  • True Star Researcher Team. Model series created by Karol Hausman, Chelsea Finn, Sergey Levine, Pete Florence and others who led Google DeepMind’s RT series.
  • Exemplary VLA + Teleop Data Approach. Well demonstrates the approach of building Robot Foundation Models with teleoperation-based data collection and VLA architecture.
  • Open-Source Friendly. Released weights and code via openpi, enabling general users to fine-tune. 1-20 hours of data is sufficient.
  • Skeptical of Humanoids/Synthetic Data. Focuses on mobile manipulators over humanoids, real data over synthetic data.
  • Must Follow. A model series worth continuous attention in the Physical AI field.

Pi Series: Robot Olympics Challenge (Physical Intelligence)

Key Significance

  • New Paradigm for VLAs: Flow Matching-based continuous action generation as an alternative to Diffusion/Autoregressive approaches
  • Real Home Environment Generalization: Pi0.5 works in completely new homes never seen during training
  • Self-Improvement via RL: Pi*0.6 uses RECAP methodology to learn from real experience
  • Fully Open-Source: Complete release of model weights, training code via openpi

Overview

Pi Series is a Vision-Language-Action model series announced by Physical Intelligence starting October 2024.

ItemDetails
CompanyPhysical Intelligence
Founded2024
GitHubPhysical-Intelligence/openpi
Blogpi.website/blog

Evolution Timeline

DateNameDescriptionDocOfficial
2024.10.31Pi0First Generalist Policy, Flow Matching + PaliGemma 3BPi0Blog
2025.01.16FAST5x faster training, DCT + BPE compressionFASTResearch
2025.02.04openpiWeights, code released (JAX + PyTorch)-Blog
2025.02.26HIRobotHuman-Robot Interaction research-Research
2025.04.22Pi0.5Open-World Generalization, Web Data Co-trainingPi0.5Blog
2025.05.28Knowledge Insulation7.5x fewer training steps-Research
2025.06.09Real-Time ChunkingReal-time control in high-latency environments-Research
2025.11.17Pi*0.6Self-improvement via RL (RECAP)Pi*0.6Blog
2025.12.16Human-to-RobotLearning from human videos-Research

Model Versions

VersionReleasedKey InnovationDetailed Doc
Pi02024.10Flow Matching VLA, 50Hz controlPi0
Pi0.52025.04Open-World generalization, Web data Co-trainingPi0.5
Pi*0.62025.11RL self-improvement, RECAPPi*0.6

Core Innovations

Flow Matching Architecture

Alternative to Diffusion, efficiently modeling complex multimodal action distributions. Generates 50Hz action chunks.

FAST Tokenizer

Detailed doc: FAST (Fast Action Tokenizer)

DCT + BPE based action compression achieving 10x compression, 5x faster training.

Knowledge Insulation

Details: pi.website/research/knowledge_insulation

Preserves VLM backbone knowledge while learning robotics. 7.5x fewer training steps via gradient blocking.

Real-Time Chunking (RTC)

Details: pi.website/research/real_time_chunking

Maintains real-time control in high-latency environments (200ms+) using inpainting approach.

RECAP

Detailed doc: Pi*0.6

RL with Experience & Corrections via Advantage-conditioned Policies. Self-improvement through demonstrations + autonomous experience + coaching.


Open Source

openpi Repository

ModelDescription
Pi0 basePretrained model for fine-tuning
Pi0-FAST baseFAST tokenizer applied version
Pi0 DROIDFranka single arm fine-tuned
Pi0 ALOHABimanual manipulation fine-tuned
  • JAX: Official implementation
  • PyTorch: HuggingFace LeRobot integration
  • Fine-tuning: 1-20 hours data, Consumer GPU capable

Research Publications

Blog Posts

DateTitle
2024.10.31Pi0: Our First Generalist Policy
2025.02.04Open Sourcing Pi0
2025.04.22Pi0.5: Open-World Generalization
2025.11.17Pi*0.6: Learning from Experience

Research

DateTitle
2025.01.16FAST: Efficient Robot Action Tokenization
2025.02.26HIRobot: Interactive Learning
2025.05.28Knowledge Insulation
2025.06.09Real-Time Action Chunking
2025.12.16Human-to-Robot Transfer

Papers

ModelLink
Pi0arXiv:2410.24164
Pi0.5arXiv:2504.16054
Pi*0.6arXiv:2511.14759

See Also