| export CUDA_VISIBLE_DEVICES=0 |
|
|
| DATASET_DIR=General-Bench-Openset |
| NLP_MODEL_NAME=Qwen/Qwen2.5-7B-Instruct |
| AUDIO_MODEL_NAME=Qwen/Qwen2-Audio-7B-Instruct |
| VIDEO_MODEL_NAME=Qwen/Qwen2.5-VL-3B-Instruct |
| IMAGE_MODEL_NAME=Qwen/Qwen2.5-VL-7B-Instruct |
| 3D_MODEL_NAME=Qwen/Qwen2.5-3B-Instruct |
|
|
| |
| STEP="123" |
| for arg in "$@"; do |
| case $arg in |
| --step=*) |
| STEP="${arg#*=}" |
| ;; |
| --step) |
| shift |
| STEP="$1" |
| ;; |
| esac |
| done |
|
|
| contains_step() { |
| case "$STEP" in |
| *$1*) return 0 ;; |
| *) return 1 ;; |
| esac |
| } |
|
|
| |
| if contains_step 1; then |
| |
| python predictors/nlp_predictor.py --dataset_dir ${DATASET_DIR}/nlp --model_name ${NLP_MODEL_NAME} |
|
|
| |
| python predictors/audio_predict_comprehension.py -m Qwen/Qwen2-Audio-7B-Instruct -d ${DATASET_DIR}/audio/comprehension/ -o ${DATASET_DIR}/audio/predictions/comprehension/ -t AccentClassification AccentSexClassification |
| python predictors/audio_predict_generation.py -m SpeechGPT -d ${DATASET_DIR}/audio/generation/ -o ${DATASET_DIR}/audio/predictions/generation/ -t SingleCaptionToAudio VideoToAudio ImageToSpeech |
|
|
| |
| python predictors/video_comprehension_tasks.py |
| python predictors/video_comprehension_flow_matching_tracking.py |
| python predictors/video_comprehension_qa_caption.py |
| python predictors/video_translation_restoration_superresolution_objectdetection.py |
| python predictors/video_generation_evaluate_kit.py |
| fi |
|
|
| MODEL_NAME=Qwen2.5-7B-Instruct |
| |
| if contains_step 2; then |
| python register.py -d ${DATASET_DIR} -t references/template_result.xlsx -o outcome -m ${MODEL_NAME} -p prediction.json |
| fi |
|
|
| MODEL_NAME=Qwen2.5-7B-Instruct |
| |
| if contains_step 3; then |
| python ranker.py -p outcome/${MODEL_NAME}_result.xlsx -m ${MODEL_NAME} |
| fi |