#!/bin/bash #SBATCH --job-name=eval-all #SBATCH --output=/home/ywan0794/MoGe/eval_all_%j.log #SBATCH --error=/home/ywan0794/MoGe/eval_all_%j.log #SBATCH --open-mode=append #SBATCH --ntasks=1 #SBATCH --cpus-per-task=4 #SBATCH --gres=gpu:H100:1 #SBATCH --time=0-12:00:00 #SBATCH --mem=80G #SBATCH --nodelist=erinyes # Single sbatch — production run for 7 models on all 10 MoGe benchmarks, serial, # one H100 held the whole time. Failures don't abort; we log & continue. # Model order: cheap → expensive (FE2E last so it doesn't block others if it crashes). export PYTHONUNBUFFERED=1 cd /home/ywan0794/MoGe source /home/ywan0794/miniconda3/etc/profile.d/conda.sh TIMESTAMP=$(date +"%Y%m%d_%H%M%S") CONFIG=/home/ywan0794/MoGe/configs/eval/all_benchmarks.json CONFIG_FE2E=/home/ywan0794/MoGe/configs/eval/fe2e_all_benchmarks.json OUT_DIR=eval_output mkdir -p $OUT_DIR SUMMARY=$OUT_DIR/_eval_all_${TIMESTAMP}.summary.txt : > $SUMMARY echo "============================================" echo "eval-all started at $(date)" echo "Config (main): $CONFIG" echo "Config (fe2e): $CONFIG_FE2E" echo "TIMESTAMP: $TIMESTAMP" echo "Summary file: $SUMMARY" echo "============================================" nvidia-smi run_model() { # Usage: run_model