--- license: apache-2.0 language: - en tags: - evaluation - video - multimodal --- # CleverHans-Evaluation Scripts and a small in-domain test set to evaluate Qwen3-Omni variants on **Video-MME**, **LVBench**, and **audio–video sync** (custom JSONL). ## What’s in the repo | Path | Purpose | |------|---------| | `setup_env.sh` | Installs **Anaconda** if conda is missing, then creates `video` (or `CONDA_ENV`) and pip-installs eval deps | | `setup_data.sh` | Downloads **all** eval data: Video-MME, LVBench, and sync eval videos + audio (to `/opt/dlami/nvme`) | | `COMMANDS.md` | Copy-paste commands: data download, merge, eval per model/dataset | | `data/kto_training_data_v2_test.jsonl` | In-domain sync test (426 lines) | | `scripts/*.py` | Download, merge, eval, metrics helpers | ## Quick start ```bash git clone https://huggingface.co/Rakancorle11/CleverHans-Evaluation cd CleverHans-Evaluation huggingface-cli login # if needed for gated models chmod +x setup_env.sh setup_data.sh bash setup_env.sh # on a machine with no conda: downloads Anaconda to ~/anaconda3 first source ~/anaconda3/etc/profile.d/conda.sh # if this is your first shell after install conda activate video bash setup_data.sh # downloads Video-MME, LVBench, sync videos + audio to /opt/dlami/nvme # Then follow COMMANDS.md — you choose which model on which benchmark. ``` **Fresh OS notes:** install `wget` before running (`sudo apt install -y wget`). System **ffmpeg** is recommended (`sudo apt install -y ffmpeg`). Override `INSTALL_DIR` / `ANACONDA_VERSION` / `CUDA_INDEX_URL` via environment variables if needed (see comments in `setup_env.sh`). ## Models (HF IDs) | Role | Model | |------|--------| | Vanilla | `Qwen/Qwen3-Omni-30B-A3B-Instruct` | | Full SFT (merge / eval base) | `Rakancorle11/qwen3omni_full_sft_revised_thinker_key` | | DPO LoRA | `Rakancorle11/Qwen3Omni-onpolicy-dpo-lora-w_audio_v2_8632`, `_v3_8632`, `_v4_8632`, `_v5_12075` | Merge LoRA into a full checkpoint for **vLLM** with `scripts/merge_adapter.py`. For **transformers-only** Video-MME/LVBench you can pass `--adapter` instead of merging. ## Data - **Video-MME / LVBench / Sync eval data**: all downloaded by `bash setup_data.sh`. - **Sync eval media** (original oops videos, random-shift videos, extracted audio): pulled from `hasnat79/ual_bench`, `Rakancorle11/random_shift_video`, `Rakancorle11/extracted_audio` into `/opt/dlami/nvme/video_source/`. ## Default paths (convention) Scripts assume a fixed split on every machine: | What | Where | |------|--------| | Benchmark videos, merged full models, sync `video_source/` (original + shifted + audio) | `/opt/dlami/nvme/...` | | Eval outputs (`eval_results.jsonl`, `metrics.json`, …) | `/home/ubuntu/eval_results/videomme`, `.../lvbench`, `.../sync` | Override with `--video-dir`, `--output-dir`, `--data-root` if your layout differs. ## Requirements - Strong GPU(s), ~200GB+ disk for benchmarks + merged weights - vLLM: `--tp` must divide **20** (audio encoder heads); e.g. `--tp 4`, not 8 - `setup_env.sh` uses CUDA 12.4 PyTorch wheels by default; override `CUDA_INDEX_URL` if needed ## Notes - Eval scripts **resume** from existing `eval_results.jsonl`. - In-domain sync: use `--data-root` so paths are not tied to `/home/ubuntu/video_source`.