| import random |
| import os |
| import argparse |
| import time |
| from vllm import LLM, SamplingParams |
| from datetime import datetime |
| from tqdm import tqdm |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| from eval_tools import apply_RL_prompt, solve_final_answer |
|
|
| from evaluate import evaluate |
| from utils import set_seed, load_jsonl, save_jsonl, construct_prompt |
| from parser import * |
| from trajectory import * |
| from data_loader import load_data |
| from python_executor import PythonExecutor |
| from model_utils import load_hf_lm_and_tokenizer, generate_completions |
| import logging |
|
|
| if not os.path.exists(f'{os.environ["modelname"]}'): |
| os.mkdir(f'{os.environ["modelname"]}') |
| if not os.path.exists(f'{os.environ["model"]}'): |
| os.mkdir(f'{os.environ["model"]}') |
|
|
| DATA_NAME = os.environ["DATA_NAME"] |
| logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(name)s - %(levelname)s - %(message)s', |
| datefmt='%Y-%m-%d %H:%M:%S', filename=f'{os.environ["model"]}/{os.environ["mode"]}-{DATA_NAME}.log', filemode='a') |
| print(f"logging in {os.environ['model']}/{os.environ['mode']}-{DATA_NAME}.log") |
|
|
| logging.info(f"modelname's infor: {os.environ['modelname']}") |
| logging.info(f"mode's infor: {os.environ['mode']}") |
| logging.info(f"model's infor: {os.environ['model']}") |
|
|
|
|
| import os |
| import multiprocessing |
| from multiprocessing import Process, Manager, Queue |
| from transformers import AutoTokenizer |
| from vllm import LLM, SamplingParams |
| import fcntl |
| import torch |
|
|
| def process_chunk(chunk, args, available_gpus, stop_words, output_queue, index, two_stage=False): |
| """ |
| 处理一个数据块的函数 |
| """ |
| GPU_LOCK_FILES = {gpu_id: f"/tmp/gpu_{gpu_id}.lock" for gpu_id in available_gpus} |
| |
| os.environ["CUDA_VISIBLE_DEVICES"] = str(available_gpus[index % len(available_gpus)]) |
| gpu_id = available_gpus[index % len(available_gpus)] |
| lock_file = f"/tmp/gpu_{gpu_id}.lock" |
|
|
| with open(lock_file, 'w') as f: |
| fcntl.flock(f, fcntl.LOCK_EX) |
|
|
| try: |
| os.environ["CUDA_VISIBLE_DEVICES"] = str(gpu_id) |
| |
| |
| llm = LLM( |
| model=args.model_name_or_path, |
| tensor_parallel_size=1, |
| pipeline_parallel_size=1, |
| trust_remote_code=True, |
| |
| gpu_memory_utilization=0.8, |
| enforce_eager=True, |
| max_seq_len_to_capture=65536, |
| ) |
|
|
| chunk_outputs = [] |
| for i in range(0, len(chunk), 125): |
| chunk_batch = chunk[i:i + 125] |
| if args.use_vllm: |
| if os.environ['stage'] == "add": |
| budget = args.max_tokens_per_call + (args.max_tokens_per_call // 50) + 50 |
| else: |
| budget = args.max_tokens_per_call + (args.max_tokens_per_call // 50) |
| os.environ["position"] = 'start' |
| chunk_batch_outputs = llm.generate( |
| chunk_batch, |
| SamplingParams( |
| temperature=args.temperature, |
| |
| top_p=0.9, |
| max_tokens=budget if not two_stage else 20, |
| n=1, |
| stop=stop_words, |
| stop_token_ids=( |
| [151645, 151643] |
| if "qwen2" in args.model_name_or_path.lower() |
| else None |
| ), |
| skip_special_tokens=False, |
| ), |
| ) |
| if os.path.exists('./start_positions.pt'): |
| os.remove('./start_positions.pt') |
| if os.path.exists('./early_positions.pt'): |
| os.remove('./early_positions.pt') |
|
|
| chunk_batch_outputs = sorted(chunk_batch_outputs, key=lambda x: int(x.request_id)) |
| chunk_batch_outputs = [output.outputs[0].text for output in chunk_batch_outputs] |
| batch_chunk = [single_chunk + chunk_output for single_chunk, chunk_output in zip(chunk_batch, chunk_batch_outputs)] |
| chunk_outputs.extend(batch_chunk) |
|
|
| output_queue.put((index, chunk_outputs)) |
| |
| del llm |
| torch.cuda.empty_cache() |
|
|
| finally: |
| if 'llm' in locals(): |
| del llm |
| |
| fcntl.flock(f, fcntl.LOCK_UN) |
|
|
|
|
|
|
| |
| def parse_args(): |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--ratio", type=float, default=-1, help="ratio of cot to use for generation") |
| parser.add_argument("--data_names", default="math", type=str) |
| parser.add_argument("--data_dir", default="./data", type=str) |
| parser.add_argument("--model_name_or_path", default="Qwen/QwQ-32B-Preview", type=str) |
| parser.add_argument("--output_dir", default="Qwen/QwQ-32B-Preview/math_eval", type=str) |
| parser.add_argument("--prompt_type", default="qwen25-math-cot", type=str) |
| parser.add_argument("--split", default="test", type=str) |
| parser.add_argument("--num_test_sample", default=-1, type=int) |
| parser.add_argument("--seed", default=0, type=int) |
| parser.add_argument("--start", default=0, type=int) |
| parser.add_argument("--end", default=-1, type=int) |
| parser.add_argument("--temperature", default=0, type=float) |
| parser.add_argument("--n_sampling", default=1, type=int) |
| parser.add_argument("--top_p", default=1, type=float) |
| parser.add_argument("--max_tokens_per_call", default=4096, type=int) |
| parser.add_argument("--shuffle", action="store_true") |
| parser.add_argument("--use_vllm", action="store_true") |
| parser.add_argument("--save_outputs", action="store_true") |
| parser.add_argument("--overwrite", action="store_true") |
| parser.add_argument("--use_safetensors", action="store_true") |
| parser.add_argument("--num_shots", type=int, default=0) |
| parser.add_argument("--apply_chat_template", action="store_true", help="Apply chat template to prompt.",) |
| parser.add_argument("--pipeline_parallel_size", type=int, default=1) |
| parser.add_argument("--adapt_few_shot", action="store_true", help="Few shot for multiple-choice questions, zero shot for others.",) |
| args = parser.parse_args() |
| args.top_p = (1 if args.temperature == 0 else args.top_p) |
| |
| |
| return args |
|
|
| def set_output_path(args, data_name): |
| |
| model_name_list = args.model_name_or_path.split('/')[-1] |
| model_name = model_name_list |
| for part in model_name_list: |
| if 'models' in part: |
| model_name = part |
| |
| |
| |
| |
| output_dir = os.path.join(args.output_dir, model_name, args.prompt_type) |
| out_file_prefix = f"{args.split}_{args.prompt_type}_{args.num_test_sample}_seed{args.seed}_t{args.temperature}" |
| out_file = f"{output_dir}/{data_name}/{out_file_prefix}_s{args.start}_e{args.end}_b{int(args.max_tokens_per_call)}_original.jsonl" |
| print(out_file) |
| os.makedirs(f"{output_dir}/{data_name}", exist_ok=True) |
| return out_file_prefix, output_dir, out_file |
|
|
|
|
| def prepare_data(data_name, args): |
| examples = load_data(data_name, args.split, args.data_dir) |
|
|
| |
| if args.num_test_sample > 0: |
| |
| examples = examples[: args.num_test_sample] |
|
|
| |
| if args.shuffle: |
| random.seed(datetime.now().timestamp()) |
| random.shuffle(examples) |
|
|
| |
| examples = examples[args.start : len(examples) if args.end == -1 else args.end] |
|
|
| |
| dt_string = datetime.now().strftime("%m-%d_%H-%M") |
| model_name = "/".join(args.model_name_or_path.split("/")[-2:]) |
| |
| |
| out_file_prefix, output_dir, out_file = set_output_path(args, data_name) |
|
|
| |
| processed_samples = [] |
| if not args.overwrite: |
| processed_files = [ |
| f |
| for f in os.listdir(f"{output_dir}/{data_name}/") |
| if f.endswith(".jsonl") and f.startswith(out_file_prefix) |
| ] |
| for f in processed_files: |
| processed_samples.extend( |
| list(load_jsonl(f"{output_dir}/{data_name}/{f}")) |
| ) |
|
|
| |
| processed_samples = {sample["idx"]: sample for sample in processed_samples} |
| processed_idxs = list(processed_samples.keys()) |
| processed_samples = list(processed_samples.values()) |
| examples = [example for example in examples if example["idx"] not in processed_idxs] |
| return examples, processed_samples, out_file |
|
|
|
|
| def setup(args): |
| |
| available_gpus = os.environ["CUDA_VISIBLE_DEVICES"].split(",") |
| if args.use_vllm: |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| llm=None |
| tokenizer = AutoTokenizer.from_pretrained( |
| args.model_name_or_path, trust_remote_code=True, max_length=16000, |
| ) |
| else: |
| llm, tokenizer = load_hf_lm_and_tokenizer( |
| model_name_or_path=args.model_name_or_path, |
| load_in_half=True, |
| use_fast_tokenizer=True, |
| use_safetensors=args.use_safetensors, |
| ) |
|
|
| |
| data_list = args.data_names.split(",") |
| results = [] |
| for data_name in data_list: |
| results.append(main(llm, tokenizer, data_name, args)) |
|
|
| |
| data_list.append("avg") |
| results.append( |
| { |
| "acc": sum([result["acc"] for result in results]) / len(results), |
| "simple_acc": sum([result["simple_acc"] for result in results]) / len(results), |
| } |
| ) |
|
|
| |
| pad = max([len(data_name) for data_name in data_list]) |
| print("\t".join(data_name.ljust(pad, " ") for data_name in data_list)) |
| print("\t".join([f"{result['acc']:.1f}".ljust(pad, " ") for result in results])) |
| |
| logging.info("\t".join(data_name.ljust(pad, " ") for data_name in data_list)) |
| logging.info(f"path = {args.model_name_or_path}") |
| logging.info(f"tip = {os.environ['tip']}") |
| logging.info(f"BUDGET = {os.environ['BUDGET']}") |
| logging.info("\t".join([f"{result['acc']:.5f}".ljust(pad, " ") for result in results])) |
| logging.info("simple_acc") |
| logging.info("\t".join([f"{result['simple_acc']:.5f}".ljust(pad, " ") for result in results])) |
|
|
|
|
| def is_multi_choice(answer): |
| for c in answer: |
| if c not in ["A", "B", "C", "D", "E"]: |
| return False |
| return True |
|
|
|
|
| def main(llm, tokenizer, data_name, args): |
| examples, processed_samples, out_file = prepare_data(data_name, args) |
| print(examples[0]) |
| print("\n" + "-" * 50) |
| print("data:", data_name, ", remain samples:", len(examples)) |
| if len(examples) > 0: |
| print(examples[0]) |
|
|
| available_gpus = os.environ["CUDA_VISIBLE_DEVICES"].split(",") |
| print(f"available_gpus = {available_gpus}") |
|
|
| |
| if "pal" in args.prompt_type: |
| executor = PythonExecutor(get_answer_expr="solution()") |
| else: |
| executor = PythonExecutor(get_answer_from_stdout=True) |
|
|
| |
| if args.ratio > 0 : |
| done_samples_path = out_file.replace("_r" + str(args.ratio), "") |
| done_samples = list(load_jsonl(done_samples_path)) |
| else: |
| done_samples = [] |
| done_samples = {sample["idx"]: sample for sample in done_samples} |
| |
| samples = [] |
| print("\nProcessing", len(examples), "examples", "=" * 50) |
| for example in tqdm(examples, total=len(examples)): |
| idx = example["idx"] |
|
|
| |
| example["question"] = parse_question(example, data_name) |
| if example["question"] == "": |
| continue |
| gt_cot, gt_ans = parse_ground_truth(example, data_name) |
| example["gt_ans"] = gt_ans |
| full_prompt = construct_prompt(example, data_name, args) |
| |
| if args.ratio > 0 : |
| done_cot = done_samples[idx]["code"][0] |
| cut_cot = done_cot[:int(len(done_cot)*args.ratio)] |
| |
| |
| |
| full_prompt = full_prompt + cut_cot + "\n\nFinal answer within \\boxed{{}}:\n" |
|
|
| |
|
|
| if idx == args.start: |
| print(full_prompt) |
|
|
| sample = { |
| "idx": idx, |
| "question": example["question"], |
| "gt_cot": gt_cot, |
| "gt": gt_ans, |
| "prompt": full_prompt, |
| } |
|
|
| |
| for key in [ |
| "level", |
| "type", |
| "unit", |
| "solution_type", |
| "choices", |
| "solution", |
| "ques_type", |
| "ans_type", |
| "answer_type", |
| "dataset", |
| "subfield", |
| "filed", |
| "theorem", |
| "answer", |
| ]: |
| if key in example: |
| sample[key] = example[key] |
| samples.append(sample) |
|
|
| |
| input_prompts = [sample["prompt"] for sample in samples for _ in range(args.n_sampling)] |
| input_prompts = apply_RL_prompt(input_prompts, args, budget = args.max_tokens_per_call) |
| |
| |
| if args.apply_chat_template: |
| input_prompts = [ |
| tokenizer.apply_chat_template( |
| [{"role": "user", "content": prompt.strip()}], |
| tokenize=False, |
| add_generation_prompt=True, |
| ) |
| for prompt in input_prompts |
| ] |
| remain_prompts = input_prompts |
| remain_prompts = [(i, prompt) for i, prompt in enumerate(remain_prompts)] |
| end_prompts = [] |
|
|
| max_func_call = 1 if args.prompt_type in ["cot", "pal", "qwen25-math-cot"] else 4 |
|
|
| stop_words = ["</s>", "<|im_end|>", "<|endoftext|>", "<|end▁of▁sentence|>"] |
|
|
| if args.prompt_type in ["cot"]: |
| stop_words.append("\n\nQuestion:") |
| if args.prompt_type in ["pal", "tool-integrated", "jiuzhang_tora"]: |
| stop_words.extend(["\n\n---", "```output"]) |
| elif args.prompt_type in ["wizard_zs", "platypus_fs"]: |
| stop_words.extend(["Instruction", "Response"]) |
| elif "jiuzhang" in args.prompt_type: |
| stop_words.append("\n\n## Question") |
| elif "numina" in args.prompt_type: |
| stop_words.append("\n### Problem") |
| elif "pure" in args.prompt_type: |
| stop_words.append("\n\n\n") |
| |
| |
| start_time = time.time() |
| print(f"start_time: {start_time}") |
| for epoch in range(max_func_call): |
| print("-" * 20, "Epoch", epoch) |
| current_prompts = remain_prompts |
| if len(current_prompts) == 0: |
| break |
|
|
| prompts = [item[1] for item in current_prompts] |
| |
|
|
| num_prompts = len(prompts) |
| chunk_size = 256 |
| outputs = [] |
|
|
| available_gpus = os.environ["CUDA_VISIBLE_DEVICES"].split(",") |
| num_gpus = len(available_gpus) |
| manager = Manager() |
| output_queue = Queue() |
| processes = [] |
| multi_outputs = [] |
| chunk_size = len(prompts) // num_gpus |
| chunks = [prompts[i:i + chunk_size] for i in range(0, len(prompts), chunk_size)] |
| num_rounds = (len(chunks) + num_gpus - 1) // num_gpus |
|
|
| for round_idx in range(num_rounds): |
| start_idx = round_idx * num_gpus |
| end_idx = min((round_idx + 1) * num_gpus, len(chunks)) |
| |
| for i in range(start_idx, end_idx): |
| chunk = chunks[i] |
| p = Process(target=process_chunk, args=(chunk, args, available_gpus, stop_words, output_queue, i)) |
| processes.append(p) |
| p.start() |
|
|
| for _ in range(len(chunks)): |
| result = output_queue.get() |
| if isinstance(result, tuple) and len(result) == 2: |
| multi_outputs.append(result) |
| else: |
| print(f"Error: Received non-tuple result: {result}") |
| |
| for p in processes: |
| p.join() |
| multi_outputs.sort(key=lambda x: x[0]) |
| outputs = [] |
| for _, chunk_output in multi_outputs: |
| outputs.extend(chunk_output) |
|
|
| print('stage one finished!!!\n' * 20) |
| |
| |
| |
| print(outputs[:3]) |
| |
| |
| |
| if os.environ['stage'] == "2": |
| print("stage 2") |
| two_stage_outputs = [] |
| modified_outputs = [] |
| print(f"len of outputs: {len(outputs)}") |
| for output in outputs: |
| |
| if "<|end▁of▁sentence|>" in output: |
| start_index = output.index("<|end▁of▁sentence|>") |
| output = output[:start_index] |
| |
| modified_output = output + "\n</think>\n\n**Final Answer**\\boxed" |
| modified_outputs.append(modified_output) |
| |
| |
| available_gpus = os.environ["CUDA_VISIBLE_DEVICES"].split(",") |
| num_gpus = len(available_gpus) |
| manager = Manager() |
| output_queue = Queue() |
| processes = [] |
| multi_outputs = [] |
| prompts = modified_outputs |
| chunk_size = len(prompts) // num_gpus |
| chunks = [prompts[i:i + chunk_size] for i in range(0, len(prompts), chunk_size)] |
| num_rounds = (len(chunks) + num_gpus - 1) // num_gpus |
|
|
| for round_idx in range(num_rounds): |
| start_idx = round_idx * num_gpus |
| end_idx = min((round_idx + 1) * num_gpus, len(chunks)) |
| |
| for i in range(start_idx, end_idx): |
| chunk = chunks[i] |
| p = Process(target=process_chunk, args=(chunk, args, available_gpus, stop_words, output_queue, i)) |
| processes.append(p) |
| p.start() |
| |
| |
| |
| |
| |
| for _ in range(len(chunks)): |
| result = output_queue.get() |
| if isinstance(result, tuple) and len(result) == 2: |
| multi_outputs.append(result) |
| else: |
| print(f"Error: Received non-tuple result: {result}") |
| |
| for p in processes: |
| p.join() |
| multi_outputs.sort(key=lambda x: x[0]) |
| outputs = [] |
| for _, chunk_output in multi_outputs: |
| outputs.extend(chunk_output) |
|
|
| elif os.environ['stage'] == "1": |
| outputs = outputs |
| |
| |
| |
| print(f"outputs:{len(outputs)}") |
| print(f"current_prompts:{len(current_prompts)}") |
| assert len(outputs) == len(current_prompts) |
|
|
| remain_prompts = [] |
| remain_codes = [] |
| for (i, query), output in zip(current_prompts, outputs): |
| output = output.rstrip() |
| query += output |
| if args.prompt_type == "pal": |
| remain_prompts.append((i, query)) |
| if "```python" in output: |
| output = extract_program(query) |
| remain_codes.append(output) |
| elif args.prompt_type == "cot": |
| end_prompts.append((i, query)) |
| elif "boxed" not in output and output.endswith("```"): |
| program = extract_program(query) |
| remain_prompts.append((i, query)) |
| remain_codes.append(program) |
| else: |
| end_prompts.append((i, query)) |
|
|
| |
| remain_results = executor.batch_apply(remain_codes) |
| for k in range(len(remain_prompts)): |
| i, query = remain_prompts[k] |
| res, report = remain_results[k] |
| exec_result = res if res else report |
| if "pal" in args.prompt_type: |
| exec_result = "\\boxed{" + exec_result + "}" |
| exec_result = f"\n```output\n{exec_result}\n```\n" |
| query += exec_result |
| |
| if epoch == max_func_call - 1: |
| query += "\nReach max function call limit." |
| remain_prompts[k] = (i, query) |
|
|
| |
| print("Unsolved samples:", len(remain_prompts)) |
| end_prompts.extend(remain_prompts) |
| |
| end_prompts = sorted(end_prompts, key=lambda x: x[0]) |
|
|
| |
| codes = [] |
| assert len(input_prompts) == len(end_prompts) |
| for i in range(len(input_prompts)): |
| if i ==1: |
| print(f"input_prompts[{i}] = {input_prompts[i]}") |
| print(f"end_prompts[{i}] = {end_prompts[i]}") |
| _, end_prompt = end_prompts[i] |
| code = end_prompt.split(input_prompts[i])[-1].strip() |
| for stop_word in stop_words: |
| if stop_word in code: |
| code = code.split(stop_word)[0].strip() |
| if args.prompt_type == "deepseek3": |
| |
| if '<|Assistant|>' in code: |
| code = code.split("<|Assistant|>")[1] |
| else: |
| code = code |
| codes.append(code) |
| |
| |
|
|
| results = [ |
| run_execute(executor, code, args.prompt_type, data_name) for code in codes |
| ] |
| time_use = time.time() - start_time |
|
|
| |
| all_samples = [] |
| for i, sample in enumerate(samples): |
| code = codes[i * args.n_sampling : (i + 1) * args.n_sampling] |
| result = results[i * args.n_sampling : (i + 1) * args.n_sampling] |
| preds = [item[0] for item in result] |
| reports = [item[1] for item in result] |
| for j in range(len(preds)): |
| if sample["gt"] in ["A", "B", "C", "D", "E"] and preds[j] not in [ |
| "A", |
| "B", |
| "C", |
| "D", |
| "E", |
| ]: |
| preds[j] = choice_answer_clean(code[j]) |
| elif is_multi_choice(sample["gt"]) and not is_multi_choice(preds[j]): |
| |
| preds[j] = "".join( |
| [c for c in preds[j] if c in ["A", "B", "C", "D", "E"]] |
| ) |
|
|
| |
| sample.update({"code": code, "pred": preds, "report": reports}) |
| all_samples.append(sample) |
|
|
| |
| all_samples.extend(processed_samples) |
| |
| all_samples, result_json = evaluate( |
| samples=all_samples, |
| data_name=data_name, |
| prompt_type=args.prompt_type, |
| execute=True, |
| ) |
|
|
| |
| if len(processed_samples) < len(all_samples) and args.save_outputs: |
| save_jsonl(all_samples, out_file) |
|
|
| result_json["time_use_in_second"] = time_use |
| result_json["time_use_in_minite"] = ( |
| f"{int(time_use // 60)}:{int(time_use % 60):02d}" |
| ) |
|
|
| with open( |
| out_file.replace(".jsonl", "_metrics.json"), "w" |
| ) as f: |
| json.dump(result_json, f, indent=4) |
| return result_json |
|
|
|
|
| if __name__ == "__main__": |
| multiprocessing.set_start_method('spawn') |
| args = parse_args() |
| set_seed(args.seed) |
| setup(args) |