FocsleFocsle
PI

π0-Distilled

64%
by Physical Intelligence

OSS distillation of π0. Validated on tabletop manipulation; HIL sim runs published alongside.

Vision-Language-ActionApache-2.0MIXEDvlamanipulationdistilled
29K downloads 640 deploymentsUpdated Apr 23, 2028
Headline:22.6ms · NVIDIA Jetson Thor · MIXED

About this model

OSS distillation of π0. Validated on tabletop manipulation; HIL sim runs published alongside.

Authored by physical-intelligence. Curated into the Fo’c’sle reference set on 2028-04-23. All cross-chip benchmarks below were collected in matched-pair runs in the HIL lab using the same input pipeline, same upstream preprocessing, and the same downstream consumer. See the methodology page for the full protocol.

Task
Vision-Language-Action
Parameters
3.2 B
Benchmarked on
1 chips
Deployments
640

Architecture

Vision-Language-Action policy
Inferred from upstream weights · simplified
RGB camsProprioGoal textVLM backboneAction headDiscretizerJoint cmds

Headline benchmarks

Training data

Pretrained on the upstream maintainer’s released checkpoint. Edge-distillation pass uses 2.4M frames from the Fo’c’sle distillation corpus (consented public data + opt-in publisher contributions). Quantization-aware fine-tune uses 320K calibration samples drawn from the target task’s eval domain.

  • Pretraining corpus: upstream maintainer release
  • Distillation corpus: 2,400,000 frames
  • Calibration set: 320,000 samples (per task)
  • Eval set: standard benchmark + matched-pair HIL runs