Chtholly17 commited on
Commit
82fb5eb
Β·
verified Β·
1 Parent(s): 3fbd09a

Update README: add author list, real GitHub/Collection/project-page URLs

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -28,9 +28,9 @@ This release is the **Stage-3 step-186** checkpoint, which gave the largest gain
28
  | Resource | Link |
29
  |---|---|
30
  | πŸ“„ Paper | <TODO_PAPER_URL> |
31
- | πŸ’» Code | https://github.com/<TODO>/VLM-CapCurriculum |
32
- | 🌐 Project page | <TODO_WEBSITE> |
33
- | πŸ€— Collection (model + data + eval) | https://huggingface.co/collections/UCSC-VLAA/<TODO> |
34
 
35
  ## Headline numbers (extended benchmark suite, AVG over 10 benchmarks)
36
 
@@ -54,7 +54,7 @@ InternVL3 needs damped optimisation in Stage 2 to avoid entropy explosion (`lr=3
54
 
55
  All three stages share **one** system / format prompt β€” see [Inference](#inference) below.
56
 
57
- Detailed launch scripts: [`training/examples/internvl3_8b/`](https://github.com/<TODO>/VLM-CapCurriculum/tree/main/training/examples/internvl3_8b) in the code repo.
58
 
59
  ## Inference
60
 
@@ -74,7 +74,7 @@ lmdeploy serve api_server UCSC-VLAA/VLM-CapCurriculum-InternVL3-8B-Staged \
74
  --server-port 23342 --tp 4
75
  ```
76
 
77
- For VLMEvalKit-style benchmark eval, plug it in via the `InternVL3_8B_Staged` alias defined in [`evaluation/configs/models.py`](https://github.com/<TODO>/VLM-CapCurriculum/blob/main/evaluation/configs/models.py).
78
 
79
  ## Intended use & limitations
80
 
@@ -89,7 +89,7 @@ Released under **Apache-2.0**, matching the upstream backbone. If you use this m
89
  ```bibtex
90
  @inproceedings{vlmcapcurriculum2026,
91
  title = {From Seeing to Thinking: Decoupling Perception and Reasoning Improves Post-Training of Vision-Language Models},
92
- author = {TODO},
93
  booktitle = {Proceedings of the International Conference on Machine Learning (ICML)},
94
  year = {2026}
95
  }
 
28
  | Resource | Link |
29
  |---|---|
30
  | πŸ“„ Paper | <TODO_PAPER_URL> |
31
+ | πŸ’» Code | https://github.com/UCSC-VLAA/VLM-CapCurriculum |
32
+ | 🌐 Project page | https://ucsc-vlaa.github.io/VLM-CapCurriculum |
33
+ | πŸ€— Collection (model + data + eval) | https://huggingface.co/collections/UCSC-VLAA/vlm-capcurriculum-from-seeing-to-thinking-icml-2026-6a07691f944148ccb2b183b8 |
34
 
35
  ## Headline numbers (extended benchmark suite, AVG over 10 benchmarks)
36
 
 
54
 
55
  All three stages share **one** system / format prompt β€” see [Inference](#inference) below.
56
 
57
+ Detailed launch scripts: [`training/examples/internvl3_8b/`](https://github.com/UCSC-VLAA/VLM-CapCurriculum/tree/main/training/examples/internvl3_8b) in the code repo.
58
 
59
  ## Inference
60
 
 
74
  --server-port 23342 --tp 4
75
  ```
76
 
77
+ For VLMEvalKit-style benchmark eval, plug it in via the `InternVL3_8B_Staged` alias defined in [`evaluation/configs/models.py`](https://github.com/UCSC-VLAA/VLM-CapCurriculum/blob/main/evaluation/configs/models.py).
78
 
79
  ## Intended use & limitations
80
 
 
89
  ```bibtex
90
  @inproceedings{vlmcapcurriculum2026,
91
  title = {From Seeing to Thinking: Decoupling Perception and Reasoning Improves Post-Training of Vision-Language Models},
92
+ author = {Juncheng Wu and Hardy Chen and Haoqin Tu and Xianfeng Tang and Freda Shi and Hui Liu and Hanqing Lu and Cihang Xie and Yuyin Zhou},
93
  booktitle = {Proceedings of the International Conference on Machine Learning (ICML)},
94
  year = {2026}
95
  }