ce-amtic commited on
Commit
ef9ad41
·
verified ·
1 Parent(s): bb1445d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -12
README.md CHANGED
@@ -19,7 +19,7 @@ base_model:
19
 
20
  # ProcVLM-2B
21
 
22
- ProcVLM-2B is a procedure-grounded vision-language model for estimating progress rewards from robot manipulation observations. Given a task description and a recent window of video frames, the model reasons about the remaining atomic actions and predicts the current task completion percentage in a `<progress>...</progress>` tag.
23
 
24
  <p align="center">
25
  <a href="https://procvlm.github.io/">Homepage</a> |
@@ -42,7 +42,6 @@ ProcVLM-2B is designed for research on robot learning, progress reward modeling,
42
 
43
  - estimating task completion progress from robot videos;
44
  - producing dense progress rewards from sparse demonstrations;
45
- - visualizing progress over time for manipulation rollouts;
46
  - adapting progress prediction to a new environment with one-shot LoRA fine-tuning.
47
 
48
  This model is not intended to be used as a safety-critical controller without downstream validation.
@@ -122,8 +121,17 @@ Given the recent observation and the task "{task}", first infer the remaining at
122
  The model should answer with reasoning and a final progress tag, for example:
123
 
124
  ```text
125
- The drawer is already open and the bread is close to the drawer. The remaining action is to place the bread inside the drawer.
126
- Therefore, the estimated progress is <progress>62.5%</progress>.
 
 
 
 
 
 
 
 
 
127
  ```
128
 
129
  ## vLLM Batch Inference
@@ -177,14 +185,14 @@ ProcVLM can be adapted to a new environment with one successful task demonstrati
177
  If you use ProcVLM, please cite the paper:
178
 
179
  ```bibtex
180
- @misc{procvlm2026,
181
- title = {ProcVLM: Learning Procedure-Grounded Progress Rewards for Robotic Manipulation},
182
- author = {ProcVLM Authors},
183
- year = {2026},
184
- eprint = {2605.08774},
185
- archivePrefix = {arXiv},
186
- primaryClass = {cs.RO},
187
- url = {https://arxiv.org/abs/2605.08774}
188
  }
189
  ```
190
 
 
19
 
20
  # ProcVLM-2B
21
 
22
+ ProcVLM-2B is a procedure-grounded vision-language model for estimating progress rewards from robot manipulation observations. Given a task description and a recent window of video frames, the model reasons about the remaining atomic actions and predicts the current task completion percentage.
23
 
24
  <p align="center">
25
  <a href="https://procvlm.github.io/">Homepage</a> |
 
42
 
43
  - estimating task completion progress from robot videos;
44
  - producing dense progress rewards from sparse demonstrations;
 
45
  - adapting progress prediction to a new environment with one-shot LoRA fine-tuning.
46
 
47
  This model is not intended to be used as a safety-critical controller without downstream validation.
 
121
  The model should answer with reasoning and a final progress tag, for example:
122
 
123
  ```text
124
+ To complete the task: Tower the blocks, the following steps are required:
125
+ 1. Grasp the green block.
126
+ 2. Place the green block onto the red block.
127
+ Therefore, the estimated progress percentage is <progress>84.13%</progress>.
128
+ ```
129
+
130
+ Or if the task is finished:
131
+
132
+ ```text
133
+ The task requires: Tower the blocks. Images show no block outside the tower, no further steps required.
134
+ Therefore, the estimated progress percentage is <progress>100.00%</progress>.
135
  ```
136
 
137
  ## vLLM Batch Inference
 
185
  If you use ProcVLM, please cite the paper:
186
 
187
  ```bibtex
188
+ @misc{feng2026procvlmlearningproceduregroundedprogress,
189
+ title={ProcVLM: Learning Procedure-Grounded Progress Rewards for Robotic Manipulation},
190
+ author={Youhe Feng and Hansen Shi and Haoyang Li and Xinlei Guo and Yang Wang and Chengyang Zhang and Jinkai Zhang and Xiaohan Zhang and Jie Tang and Jing Zhang},
191
+ year={2026},
192
+ eprint={2605.08774},
193
+ archivePrefix={arXiv},
194
+ primaryClass={cs.RO},
195
+ url={https://arxiv.org/abs/2605.08774},
196
  }
197
  ```
198