| import json |
| from transformers import AutoTokenizer |
| from typing import Any |
| import numpy as np |
|
|
| def convert_data_to_id(tokenizer: AutoTokenizer, data: Any): |
| input_ids = tokenizer.encode(data) |
| ids = input_ids |
| ids = np.array(ids, dtype=np.int32) |
| return ids |
|
|
| def get_tokenizer(tokenizer_path): |
| tokenizer = AutoTokenizer.from_pretrained( |
| tokenizer_path, use_fast=not False, trust_remote_code=False |
| ) |
| return tokenizer |
|
|
| |
| source_file = "../redstone_v4_23_json/mix_splits/mixed_redstone_part_20.jsonl" |
| out_file = "256k_docs_for_test_qwen.jsonl" |
| tokenizer_path = "../Qwen2.5-1.5B" |
| min_len = 256*1024 |
| retri_num = 1000 |
|
|
| tokenizer = get_tokenizer(tokenizer_path) |
| idx = 0 |
| succ_cnt = 0 |
| out_f = open(out_file,'w') |
|
|
| with open(source_file) as f: |
| for line in f: |
| idx += 1 |
| if idx % 10000 == 0: |
| print('Cur idx - ', idx) |
| line = json.loads(line) |
| cur_texts = [] |
| if 'text' in line: |
| temp = line['text'] |
| elif 'raw_content_lines' in line: |
| temp = "\n".join(line['raw_content_lines']) |
| else: |
| print("error") |
| exit() |
| try: |
| token_id = convert_data_to_id(tokenizer, temp) |
| except UnicodeDecodeError: |
| print('Error line - encoding: ', idx) |
| if len(token_id) > min_len: |
| temp_dic = {'text': temp} |
| out_f.write(json.dumps(temp_dic) +"\n") |
| succ_cnt += 1 |
| if succ_cnt % 10==0: |
| print("succ_cnt:",succ_cnt) |
| if succ_cnt==1000: |
| break |
| out_f.close() |
| print(f"retrieve {succ_cnt} docs longer than {min_len} from {idx} docs.") |