Update README.md
Browse files
README.md
CHANGED
|
@@ -19,7 +19,7 @@ base_model:
|
|
| 19 |
|
| 20 |
# ProcVLM-2B
|
| 21 |
|
| 22 |
-
ProcVLM-2B is a procedure-grounded vision-language model for estimating progress rewards from robot manipulation observations. Given a task description and a recent window of video frames, the model reasons about the remaining atomic actions and predicts the current task completion percentage
|
| 23 |
|
| 24 |
<p align="center">
|
| 25 |
<a href="https://procvlm.github.io/">Homepage</a> |
|
|
@@ -42,7 +42,6 @@ ProcVLM-2B is designed for research on robot learning, progress reward modeling,
|
|
| 42 |
|
| 43 |
- estimating task completion progress from robot videos;
|
| 44 |
- producing dense progress rewards from sparse demonstrations;
|
| 45 |
-
- visualizing progress over time for manipulation rollouts;
|
| 46 |
- adapting progress prediction to a new environment with one-shot LoRA fine-tuning.
|
| 47 |
|
| 48 |
This model is not intended to be used as a safety-critical controller without downstream validation.
|
|
@@ -122,8 +121,17 @@ Given the recent observation and the task "{task}", first infer the remaining at
|
|
| 122 |
The model should answer with reasoning and a final progress tag, for example:
|
| 123 |
|
| 124 |
```text
|
| 125 |
-
|
| 126 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 127 |
```
|
| 128 |
|
| 129 |
## vLLM Batch Inference
|
|
@@ -177,14 +185,14 @@ ProcVLM can be adapted to a new environment with one successful task demonstrati
|
|
| 177 |
If you use ProcVLM, please cite the paper:
|
| 178 |
|
| 179 |
```bibtex
|
| 180 |
-
@misc{
|
| 181 |
-
|
| 182 |
-
|
| 183 |
-
|
| 184 |
-
|
| 185 |
-
|
| 186 |
-
|
| 187 |
-
|
| 188 |
}
|
| 189 |
```
|
| 190 |
|
|
|
|
| 19 |
|
| 20 |
# ProcVLM-2B
|
| 21 |
|
| 22 |
+
ProcVLM-2B is a procedure-grounded vision-language model for estimating progress rewards from robot manipulation observations. Given a task description and a recent window of video frames, the model reasons about the remaining atomic actions and predicts the current task completion percentage.
|
| 23 |
|
| 24 |
<p align="center">
|
| 25 |
<a href="https://procvlm.github.io/">Homepage</a> |
|
|
|
|
| 42 |
|
| 43 |
- estimating task completion progress from robot videos;
|
| 44 |
- producing dense progress rewards from sparse demonstrations;
|
|
|
|
| 45 |
- adapting progress prediction to a new environment with one-shot LoRA fine-tuning.
|
| 46 |
|
| 47 |
This model is not intended to be used as a safety-critical controller without downstream validation.
|
|
|
|
| 121 |
The model should answer with reasoning and a final progress tag, for example:
|
| 122 |
|
| 123 |
```text
|
| 124 |
+
To complete the task: Tower the blocks, the following steps are required:
|
| 125 |
+
1. Grasp the green block.
|
| 126 |
+
2. Place the green block onto the red block.
|
| 127 |
+
Therefore, the estimated progress percentage is <progress>84.13%</progress>.
|
| 128 |
+
```
|
| 129 |
+
|
| 130 |
+
Or if the task is finished:
|
| 131 |
+
|
| 132 |
+
```text
|
| 133 |
+
The task requires: Tower the blocks. Images show no block outside the tower, no further steps required.
|
| 134 |
+
Therefore, the estimated progress percentage is <progress>100.00%</progress>.
|
| 135 |
```
|
| 136 |
|
| 137 |
## vLLM Batch Inference
|
|
|
|
| 185 |
If you use ProcVLM, please cite the paper:
|
| 186 |
|
| 187 |
```bibtex
|
| 188 |
+
@misc{feng2026procvlmlearningproceduregroundedprogress,
|
| 189 |
+
title={ProcVLM: Learning Procedure-Grounded Progress Rewards for Robotic Manipulation},
|
| 190 |
+
author={Youhe Feng and Hansen Shi and Haoyang Li and Xinlei Guo and Yang Wang and Chengyang Zhang and Jinkai Zhang and Xiaohan Zhang and Jie Tang and Jing Zhang},
|
| 191 |
+
year={2026},
|
| 192 |
+
eprint={2605.08774},
|
| 193 |
+
archivePrefix={arXiv},
|
| 194 |
+
primaryClass={cs.RO},
|
| 195 |
+
url={https://arxiv.org/abs/2605.08774},
|
| 196 |
}
|
| 197 |
```
|
| 198 |
|