| MODEL_ROOT="out/pretrain/tinyllama_lr_plus" | |
| #"out/pretrain/tinyllama_3_epoch" | |
| #"out/pretrain/qwen2" | |
| #"out/pretrain/tinyllama" | |
| TASKS="arxiv_mc" | |
| BATCH=4 | |
| OUT_ROOT="out/eval/tinyllama_lr_plus_arxiv_mc" | |
| #"out/eval/tinyllama_3_epoch_arxiv_mc" | |
| #"out/eval/qwen2_arxiv_mc" | |
| #"out/eval/tinyllama_arxiv_mc" | |
| MONTHS=(2407 2408 2409 2410 2411 2412 2501 2502 2503 2504 2505 2506) | |
| for ((i=0; i<${#MONTHS[@]}; i+=1)); do | |
| m=${MONTHS[$i]} | |
| ckpt="${MODEL_ROOT}/${m}/final" | |
| out="${OUT_ROOT}/${m}" | |
| mkdir -p "$out" | |
| litgpt evaluate "${ckpt}" \ | |
| --batch_size ${BATCH} \ | |
| --tasks "${TASKS}" \ | |
| --out_dir "${out}" \ | |
| 2>&1 | tee "${out}/log.txt" | |
| done | |
| echo "All months evaluated." |