| --- |
| license: apache-2.0 |
| library_name: vla-foundry |
| tags: |
| - foundry |
| - vla_foundry |
| - vla |
| - robotics |
| - diffusion-policy |
| - flow-matching |
| pipeline_tag: robotics |
| --- |
| |
| # Foundry-VLA-1.7B-sim |
|
|
| A 1.7B parameter vision-language-action model for bimanual robotic manipulation, part of the [VLA Foundry](https://github.com/TRI-ML/vla_foundry) collection. Trained on simulated manipulation data only. |
|
|
| ## Model Description |
|
|
| - **Architecture:** Foundry-VLM-1.3B vision-language backbone + (condition on last 4 layers) flow-matching diffusion action head (24 layers, 1024 hidden dim, 16 heads) |
| - **Parameters:** 1.7B (non-embedding) |
| - **Action space:** 20-dim relative actions (bimanual xyz + 6D rotation + gripper) |
| - **Cameras:** 4 views (2 scene + 2 wrist) |
| - **Training data:** 102M samples from simulated bimanual manipulation tasks only |
| - **VLM backbone:** [Foundry-VLM-1.3B-200M](https://huggingface.co/TRI-ML/Foundry-VLM-1.3B-200M) |
|
|
| ## Evaluation Results |
|
|
| Success rates on 16 seen tasks and 3 unseen tasks (200 rollouts per task): |
|
|
| | Simulator | Seen (16 tasks) | Unseen (3 tasks) | |
| |---|---|---| |
| | CS | 60.3% | 8.2% | |
| | OSS | 41.0% | 11.7% | |
|
|
| ## Usage |
|
|
| ```bash |
| git clone https://github.com/TRI-ML/vla_foundry.git |
| cd vla_foundry |
| pip install -e . |
| ``` |
|
|
| ```python |
| from vla_foundry.models.base_model import BaseModel |
| model = BaseModel.from_pretrained("TRI-ML/Foundry-VLA-1.7B-sim") |
| ``` |
|
|
| ## Links |
|
|
| - **Project page:** [tri-ml.github.io/vla_foundry](https://tri-ml.github.io/vla_foundry/) |
| - **Paper:** [VLA Foundry (arXiv 2604.19728)](https://arxiv.org/abs/2604.19728) |
| - **Code:** [github.com/TRI-ML/vla_foundry](https://github.com/TRI-ML/vla_foundry) |
| - **Collection:** [VLA Foundry collection](https://huggingface.co/collections/TRI-ML/vla-foundry) |
|
|