# SGL-new This repository is a cleaned, submission-oriented copy of the SGL codebase for TextVQA large-only experiments: 1. `InternVL2-2B` large-only 2. `InternVL2-8B` large-only 3. `InternVL2-26B` large-only 4. `2B vision + 1B mlp1 + 1B language model` hybrid checkpoint large-only 5. `2B vision + 8B mlp1 + 8B language model` hybrid checkpoint large-only 6. `2B vision + 26B mlp1 + 26B language model` hybrid checkpoint large-only The repository does **not** include checkpoints or datasets. The intended workflow is: 1. create an environment 2. place checkpoints under `checkpoints/` 3. prepare TextVQA data under `data/` 4. optionally build the hybrid checkpoint 5. run one of the experiment launch scripts ## 1. Repository Structure Main experiment scripts: - `textvqa2B-largeonly.sh` - `textvqa8B-largeonly.sh` - `textvqa26B-largeonly.sh` - `textvqaHybrid-2Bvision-1Bllm-largeonly.sh` - `textvqaHybrid-2Bvision-8Bllm-largeonly.sh` - `textvqaHybrid-2Bvision-26Bllm-largeonly.sh` - `run_textvqa_three_largeonly.sh` - `run_textvqa_five_largeonly.sh` - `train_textvqaHybrid-2Bvision-26Bllm-mlp.sh` Core evaluation code: - `eval/vqa/run_single_model_native.py` Native single-model helpers: - `eval/vqa/run_single_model_native.py` - `eval/vqa/run_full_textvqa_native.sh` Utility scripts: - `tools/prepare_textvqa_for_sgl.py` - `tools/build_hybrid_checkpoint.py` - `build_hybrid_checkpoint_2bvision_1bllm.sh` - `tools/hybrid_single_infer.py` - `tools/train_hybrid_textvqa_mlp.py` - `build_hybrid_checkpoint_2bvision_26bllm.sh` Environment helper: - `setup_sgl_2b_env.sh` ## 2. Environment Setup This repo expects Python 3.10 and a CUDA-enabled PyTorch installation. ### Option A: manual setup ```bash conda create -y -n sgl-new python=3.10 conda activate sgl-new pip install --upgrade pip # Install torch/torchvision matching your CUDA version. # Example for CUDA 12.1: pip install torch torchvision --index-url https://download.pytorch.org/whl/cu121 pip install -r requirements.txt ``` ### Option B: helper script ```bash bash setup_sgl_2b_env.sh sgl-new conda activate sgl-new # Then install torch/torchvision matching your CUDA version. pip install torch torchvision --index-url https://download.pytorch.org/whl/cu121 ``` ### Notes - `flash-attn` is optional. The code can run without it, but may be slower. - The large-only launchers now call Python directly and optionally shard a model with `device_map`. - If `transformers` or `torch` versions are changed substantially, verify that `InternVL` remote-code loading still works. ## 3. Checkpoint Layout Create a directory: ```bash mkdir -p checkpoints ``` Place checkpoints under `checkpoints/` with these names: - `checkpoints/models--OpenGVLab--InternVL2-1B` - `checkpoints/models--OpenGVLab--InternVL2-2B` - `checkpoints/models--OpenGVLab--InternVL2-8B` - `checkpoints/models--OpenGVLab--InternVL2-26B` The hybrid checkpoint will be created at: - `checkpoints/InternVL2-1B_2Bvision_hybrid` - `checkpoints/InternVL2-8B_2Bvision_hybrid` - `checkpoints/InternVL2-26B_2Bvision_hybrid` If you want to use a different checkpoint layout, override `CHECKPOINT_ROOT` or `CHECKPOINT` when launching. ## 4. TextVQA Data Preparation This repo expects SGL-style TextVQA files under: - `data/textvqa/textvqa_train.jsonl` - `data/textvqa/textvqa_val.jsonl` - `data/textvqa/textvqa_val_questions.json` - `data/textvqa/textvqa_val_annotations.json` The repo does **not** ship the dataset. ### 4.1 Download the official TextVQA data Prepare: - `TextVQA_0.5.1_train.json` - `TextVQA_0.5.1_val.json` - `TextVQA_0.5.1_test.json` - training/validation images - test images Place them under: ```text data/textvqa_official/ ├── TextVQA_0.5.1_train.json ├── TextVQA_0.5.1_val.json ├── TextVQA_0.5.1_test.json ├── train_images/ └── test_images/ ``` ### 4.2 Convert official data to SGL format From the repo root: ```bash python tools/prepare_textvqa_for_sgl.py \ --official-root data/textvqa_official \ --output-root data/textvqa ``` This script: - creates `data/textvqa/*.jsonl` - creates `textvqa_val_questions.json` - creates `textvqa_val_annotations.json` - symlinks `train_images` and `test_images` into `data/textvqa/` ## 5. Building Hybrid Checkpoints ### 5.1 2B vision + 1B LLM hybrid The hybrid experiment means: - `vision_model` from `InternVL2-2B` - `mlp1` from `InternVL2-1B` - `language_model` from `InternVL2-1B` Use the convenience wrapper: ```bash bash build_hybrid_checkpoint_2bvision_1bllm.sh ``` Equivalent manual command: ```bash python tools/build_hybrid_checkpoint.py \ --base-checkpoint checkpoints/models--OpenGVLab--InternVL2-1B \ --vision-checkpoint checkpoints/models--OpenGVLab--InternVL2-2B \ --output-dir checkpoints/InternVL2-1B_2Bvision_hybrid ``` ### 5.2 2B vision + 8B LLM hybrid The hybrid experiment means: - `vision_model` from `InternVL2-2B` - `mlp1` from `InternVL2-8B` - `language_model` from `InternVL2-8B` In this repo, the reproducible builder is: - `tools/build_hybrid_checkpoint.py` Run: ```bash python tools/build_hybrid_checkpoint.py \ --base-checkpoint checkpoints/models--OpenGVLab--InternVL2-8B \ --vision-checkpoint checkpoints/models--OpenGVLab--InternVL2-2B \ --output-dir checkpoints/InternVL2-8B_2Bvision_hybrid ``` This script starts from the 8B checkpoint, replaces its `vision_model` weights with the 2B `vision_model`, and saves a new merged checkpoint. ### 5.3 2B vision + 26B LLM hybrid Use the convenience wrapper: ```bash bash build_hybrid_checkpoint_2bvision_26bllm.sh ``` Equivalent manual command: ```bash python tools/build_hybrid_checkpoint.py \ --base-checkpoint checkpoints/models--OpenGVLab--InternVL2-26B \ --vision-checkpoint checkpoints/models--OpenGVLab--InternVL2-2B \ --output-dir checkpoints/InternVL2-26B_2Bvision_hybrid ``` ## 6. How the Experiments Map to Code ### 6.1 InternVL2-2B large-only Launcher: - `textvqa2B-largeonly.sh` Core code path: - `eval/vqa/run_single_model_native.py --mode textvqa_eval` Default checkpoint: - `checkpoints/models--OpenGVLab--InternVL2-2B` Run: ```bash bash textvqa2B-largeonly.sh ``` Optional overrides: ```bash CHECKPOINT_ROOT=/path/to/checkpoints \ OUT_DIR=/path/to/output \ GPUS_PER_MODEL=1 \ bash textvqa2B-largeonly.sh ``` ### 6.2 InternVL2-8B large-only Launcher: - `textvqa8B-largeonly.sh` Core code path: - `eval/vqa/run_single_model_native.py --mode textvqa_eval` Default checkpoint: - `checkpoints/models--OpenGVLab--InternVL2-8B` Run: ```bash bash textvqa8B-largeonly.sh ``` Optional overrides: ```bash CHECKPOINT_ROOT=/path/to/checkpoints \ OUT_DIR=/path/to/output \ GPUS_PER_MODEL=1 \ bash textvqa8B-largeonly.sh ``` ### 6.3 InternVL2-26B large-only Launcher: - `textvqa26B-largeonly.sh` Core code path: - `eval/vqa/run_single_model_native.py --mode textvqa_eval` Default checkpoint: - `checkpoints/models--OpenGVLab--InternVL2-26B` Run: ```bash bash textvqa26B-largeonly.sh ``` Optional overrides: ```bash CUDA_VISIBLE_DEVICES=0,1 \ CHECKPOINT_ROOT=/path/to/checkpoints \ OUT_DIR=/path/to/output \ GPUS_PER_MODEL=2 \ bash textvqa26B-largeonly.sh ``` ### 6.4 2B vision + 1B mlp1 + 1B language model large-only Launcher: - `textvqaHybrid-2Bvision-1Bllm-largeonly.sh` Core code path: - `eval/vqa/run_single_model_native.py --mode textvqa_eval` Hybrid builder: - `build_hybrid_checkpoint_2bvision_1bllm.sh` - `tools/build_hybrid_checkpoint.py` Default checkpoint: - `checkpoints/InternVL2-1B_2Bvision_hybrid` Run: ```bash bash textvqaHybrid-2Bvision-1Bllm-largeonly.sh ``` Optional overrides: ```bash CHECKPOINT_ROOT=/path/to/checkpoints \ OUT_DIR=/path/to/output \ GPUS_PER_MODEL=1 \ bash textvqaHybrid-2Bvision-1Bllm-largeonly.sh ``` ### 6.5 2B vision + 8B mlp1 + 8B language model large-only Launcher: - `textvqaHybrid-2Bvision-8Bllm-largeonly.sh` Core code path: - `eval/vqa/run_single_model_native.py --mode textvqa_eval` Hybrid builder: - `tools/build_hybrid_checkpoint.py` Default checkpoint: - `checkpoints/InternVL2-8B_2Bvision_hybrid` Run: ```bash bash textvqaHybrid-2Bvision-8Bllm-largeonly.sh ``` Optional overrides: ```bash CHECKPOINT_ROOT=/path/to/checkpoints \ OUT_DIR=/path/to/output \ GPUS_PER_MODEL=1 \ bash textvqaHybrid-2Bvision-8Bllm-largeonly.sh ``` ### 6.6 2B vision + 26B mlp1 + 26B language model large-only Launcher: - `textvqaHybrid-2Bvision-26Bllm-largeonly.sh` Core code path: - `eval/vqa/run_single_model_native.py --mode textvqa_eval` Hybrid builder: - `build_hybrid_checkpoint_2bvision_26bllm.sh` - `tools/build_hybrid_checkpoint.py` Default checkpoint: - `checkpoints/InternVL2-26B_2Bvision_hybrid` Run: ```bash bash textvqaHybrid-2Bvision-26Bllm-largeonly.sh ``` Optional overrides: ```bash CUDA_VISIBLE_DEVICES=0,1 \ CHECKPOINT_ROOT=/path/to/checkpoints \ OUT_DIR=/path/to/output \ GPUS_PER_MODEL=2 \ bash textvqaHybrid-2Bvision-26Bllm-largeonly.sh ``` ### 6.7 Optional CoT-style reasoning The native and hybrid inference entry points now support optional reasoning modes: - `--reasoning-mode none`: default single-pass decoding - `--reasoning-mode prompt`: adds an internal "think step by step" instruction in one pass - `--reasoning-mode two_pass`: first generates explicit reasoning, then compresses it into the final short answer If you do not set `REASONING_MODE` or `--reasoning-mode`, the code stays on the original normal inference path. For the hybrid TextVQA launchers, use environment variables: ```bash REASONING_MODE=two_pass \ REASONING_MAX_NEW_TOKENS=64 \ SAVE_REASONING=1 \ bash textvqaHybrid-2Bvision-8Bllm-largeonly.sh ``` For the shared-vision launcher: ```bash REASONING_MODE=two_pass \ REASONING_MAX_NEW_TOKENS=64 \ SAVE_REASONING=1 \ bash textvqaSharedVision-2Bguide-8Btext.sh ``` To let the small guide model produce a short text hint for the large decoder: ```bash GUIDE_TEXT_MODE=short_rationale \ GUIDE_TEXT_MAX_NEW_TOKENS=12 \ bash textvqaSharedVision-2Bguide-8Btext.sh ``` To force a short CoT on the guide branch so its generation changes the visual-token attention scores: ```bash GUIDE_REASONING_MODE=short_cot \ GUIDE_REASONING_MAX_NEW_TOKENS=1024 \ bash textvqaSharedVision-2Bguide-8Btext.sh ``` Both options can be enabled together. For single-image hybrid debugging: ```bash python tools/hybrid_single_infer.py \ --vision-checkpoint checkpoints/models--OpenGVLab--InternVL2-2B \ --language-checkpoint checkpoints/models--OpenGVLab--InternVL2-8B \ --image-path /path/to/image.jpg \ --prompt "What is the brand name on the sign?" \ --reasoning-mode two_pass \ --reasoning-max-new-tokens 64 \ --answer-format-prompt "Answer the question using a single word or phrase." ``` ## 7. Running Sequential Launchers Use: ```bash bash run_textvqa_three_largeonly.sh ``` Default output root: - `outputs/textvqa_three_largeonly` This script runs: 1. 2B 2. 8B 3. hybrid 2B-vision + 8B-LLM each with its own output subdirectory and launcher log. To run all five experiments, use: ```bash bash run_textvqa_five_largeonly.sh ``` This script adds: 1. 26B 2. hybrid 2B-vision + 26B-LLM ## 8. Minimal Hybrid Fine-Tuning On TextVQA For a lightweight experiment, this repo also includes a minimal script that: 1. builds `2B vision + 26B mlp1 + 26B language_model` 2. freezes everything except `mlp1` 3. trains on TextVQA jsonl 4. runs validation inference immediately after training Launcher: - `train_textvqaHybrid-2Bvision-26Bllm-mlp.sh` Core code: - `tools/train_hybrid_textvqa_mlp.py` Default demo dataset: - `/home/yf/snap/SGL_yf/data/textvqa_demo_backup/textvqa_train.jsonl` - `/home/yf/snap/SGL_yf/data/textvqa_demo_backup/textvqa_val.jsonl` Run: ```bash bash train_textvqaHybrid-2Bvision-26Bllm-mlp.sh ``` Important assumptions: - `UPSTREAM_SGL_ROOT` defaults to `/home/yf/snap/SGL` because this script reuses the upstream `internvl` package. - The default launcher expects local checkpoints at: - `/root/model_ckpts/models--OpenGVLab--InternVL2-2B` - `/root/model_ckpts/models--OpenGVLab--InternVL2-26B` - The minimal implementation currently supports `batch_size=1`. ## 9. Native Single-Model Inference Utilities These are not required for the main large-only experiments, but they are included because they are useful for debugging and single-sample inspection. ### Single sample or single question Code: - `eval/vqa/run_single_model_native.py` Example: ```bash python eval/vqa/run_single_model_native.py \ --checkpoint checkpoints/models--OpenGVLab--InternVL2-2B \ --mode single \ --image-path /path/to/image.jpg \ --prompt "What is written on the sign?" \ --max-new-tokens 32 \ --dynamic ``` ### Full TextVQA native evaluation for 2B and 8B Code: - `eval/vqa/run_full_textvqa_native.sh` Example: ```bash bash eval/vqa/run_full_textvqa_native.sh outputs/native_eval ``` ## 10. Hybrid Single-Sample Debugging Utility Code: - `tools/hybrid_single_infer.py` Example: ```bash python tools/hybrid_single_infer.py \ --vision-checkpoint checkpoints/models--OpenGVLab--InternVL2-2B \ --language-checkpoint checkpoints/models--OpenGVLab--InternVL2-8B \ --image-path /path/to/image.jpg \ --prompt "What is written on the sign?" \ --dynamic ``` This script does **not** require a saved hybrid checkpoint. It builds the hybrid model in memory for single-sample inspection. ## 11. Output Files The large-only evaluation script writes outputs under the launcher-provided output directory. Typical files include one JSON results file per run inside the launcher-provided output directory. ## 12. Minimal Reproduction Checklist For someone receiving this repository, the minimal steps are: 1. create a Python environment 2. install `torch`, `torchvision`, and `requirements.txt` 3. download `InternVL2-2B`, `InternVL2-8B`, and optionally `InternVL2-26B` into `checkpoints/` 4. download official TextVQA into `data/textvqa_official/` 5. run `python tools/prepare_textvqa_for_sgl.py` 6. run `python tools/build_hybrid_checkpoint.py` 7. run one of: - `bash textvqa2B-largeonly.sh` - `bash textvqa8B-largeonly.sh` - `bash textvqa26B-largeonly.sh` - `bash textvqaHybrid-2Bvision-8Bllm-largeonly.sh` - `bash textvqaHybrid-2Bvision-26Bllm-largeonly.sh` ## 13. Important Assumptions - The code assumes CUDA is available for model inference. - The code assumes TextVQA data is prepared under `data/textvqa/`. - The code assumes checkpoints are available under `checkpoints/` unless overridden. - All large-only experiments use the same evaluation implementation: `eval/vqa/run_single_model_native.py --mode textvqa_eval` - `InternVL2-26B` and the `2B vision + 26B LLM` hybrid usually require multiple visible GPUs.