diff --git "a/bn/baseline/data_15000_1000/train.log" "b/bn/baseline/data_15000_1000/train.log" new file mode 100644--- /dev/null +++ "b/bn/baseline/data_15000_1000/train.log" @@ -0,0 +1,2145 @@ +W0626 23:48:19.327176 1412956 site-packages/torch/distributed/run.py:766] +W0626 23:48:19.327176 1412956 site-packages/torch/distributed/run.py:766] ***************************************** +W0626 23:48:19.327176 1412956 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. +W0626 23:48:19.327176 1412956 site-packages/torch/distributed/run.py:766] ***************************************** +[2025-06-26 23:48:25,050] [INFO] [real_accelerator.py:254:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-06-26 23:48:25,125] [INFO] [real_accelerator.py:254:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-06-26 23:48:25,167] [INFO] [real_accelerator.py:254:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-06-26 23:48:25,168] [INFO] [real_accelerator.py:254:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-06-26 23:48:26,605] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False +[2025-06-26 23:48:26,715] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False +[2025-06-26 23:48:26,733] [INFO] [comm.py:675:init_distributed] cdb=None +[2025-06-26 23:48:26,733] [INFO] [comm.py:706:init_distributed] Initializing TorchBackend in DeepSpeed with backend nccl +[2025-06-26 23:48:26,748] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False +[2025-06-26 23:48:26,778] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False +[2025-06-26 23:48:26,847] [INFO] [comm.py:675:init_distributed] cdb=None +[2025-06-26 23:48:26,880] [INFO] [comm.py:675:init_distributed] cdb=None +[2025-06-26 23:48:26,915] [INFO] [comm.py:675:init_distributed] cdb=None +06/26/2025 23:48:27 - WARNING - __main__ - Process rank: 0, device: cuda:0, n_gpu: 1distributed training: True, 16-bits training: False +06/26/2025 23:48:27 - INFO - __main__ - Training/evaluation parameters LoRATrainingArguments( +_n_gpu=1, +accelerator_config={'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None, 'use_configured_state': False}, +adafactor=False, +adam_beta1=0.9, +adam_beta2=0.999, +adam_epsilon=1e-08, +auto_find_batch_size=False, +batch_eval_metrics=False, +bf16=True, +bf16_full_eval=True, +data_seed=None, +dataloader_drop_last=False, +dataloader_num_workers=2, +dataloader_persistent_workers=False, +dataloader_pin_memory=True, +dataloader_prefetch_factor=None, +ddp_backend=None, +ddp_broadcast_buffers=None, +ddp_bucket_cap_mb=None, +ddp_find_unused_parameters=None, +ddp_timeout=3600, +debug=[], +deepspeed=./config/deepspeed_config.json, +disable_tqdm=False, +dispatch_batches=None, +do_eval=True, +do_predict=False, +do_train=True, +eval_accumulation_steps=None, +eval_delay=0, +eval_do_concat_batches=True, +eval_on_start=True, +eval_steps=200, +eval_strategy=steps, +eval_use_gather_object=False, +evaluation_strategy=None, +fp16=False, +fp16_backend=auto, +fp16_full_eval=False, +fp16_opt_level=O1, +fsdp=[], +fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, +fsdp_min_num_params=0, +fsdp_transformer_layer_cls_to_wrap=None, +full_determinism=False, +gradient_accumulation_steps=1, +gradient_checkpointing=True, +gradient_checkpointing_kwargs=None, +greater_is_better=False, +group_by_length=False, +half_precision_backend=auto, +hub_always_push=False, +hub_model_id=None, +hub_private_repo=False, +hub_strategy=every_save, +hub_token=, +ignore_data_skip=False, +include_inputs_for_metrics=False, +include_num_input_tokens_seen=False, +include_tokens_per_second=False, +jit_mode_eval=False, +label_names=None, +label_smoothing_factor=0.0, +learning_rate=0.0005, +length_column_name=length, +load_best_model_at_end=True, +load_lora_from=None, +local_rank=0, +log_level=passive, +log_level_replica=warning, +log_on_each_node=True, +logging_dir=./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/runs/Jun26_23-48-24_innmi1srh2-p040, +logging_first_step=False, +logging_nan_inf_filter=True, +logging_steps=1.0, +logging_strategy=steps, +lora_config=./config/lora_config.json, +lr_scheduler_kwargs={}, +lr_scheduler_type=inverse_sqrt, +max_grad_norm=1.0, +max_steps=-1, +metric_for_best_model=eval_loss, +mp_parameters=, +neftune_noise_alpha=None, +no_cuda=False, +num_train_epochs=5.0, +optim=adamw_torch, +optim_args=None, +optim_target_modules=None, +output_dir=./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/, +overwrite_output_dir=True, +past_index=-1, +per_device_eval_batch_size=25, +per_device_train_batch_size=25, +prediction_loss_only=False, +push_to_hub=False, +push_to_hub_model_id=None, +push_to_hub_organization=None, +push_to_hub_token=, +ray_scope=last, +remove_unused_columns=True, +report_to=['wandb'], +restore_callback_states_from_checkpoint=False, +resume_from_checkpoint=None, +run_name=./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/, +save_on_each_node=False, +save_only_model=False, +save_safetensors=True, +save_steps=200, +save_strategy=steps, +save_total_limit=1, +seed=1, +skip_memory_metrics=True, +split_batches=None, +tf32=None, +torch_compile=False, +torch_compile_backend=None, +torch_compile_mode=None, +torch_empty_cache_steps=None, +torchdynamo=None, +tpu_metrics_debug=False, +tpu_num_cores=None, +use_cpu=False, +use_int8_training=False, +use_ipex=False, +use_legacy_prediction_loop=False, +use_lora=True, +use_mps_device=False, +warmup_ratio=0.03, +warmup_steps=0, +weight_decay=0.0, +) +06/26/2025 23:48:27 - WARNING - __main__ - Process rank: 3, device: cuda:3, n_gpu: 1distributed training: True, 16-bits training: False +06/26/2025 23:48:27 - WARNING - __main__ - Process rank: 2, device: cuda:2, n_gpu: 1distributed training: True, 16-bits training: False +06/26/2025 23:48:27 - WARNING - __main__ - Process rank: 1, device: cuda:1, n_gpu: 1distributed training: True, 16-bits training: False +Using custom data configuration default-10eaf7c5c1c6f11a +06/26/2025 23:48:27 - INFO - datasets.builder - Using custom data configuration default-10eaf7c5c1c6f11a +Loading Dataset Infos from /home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/datasets/packaged_modules/json +06/26/2025 23:48:27 - INFO - datasets.info - Loading Dataset Infos from /home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/datasets/packaged_modules/json +Overwrite dataset info from restored data version if exists. +06/26/2025 23:48:27 - INFO - datasets.builder - Overwrite dataset info from restored data version if exists. +Loading Dataset info from /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092 +06/26/2025 23:48:27 - INFO - datasets.info - Loading Dataset info from /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092 +Found cached dataset json (/home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092) +06/26/2025 23:48:27 - INFO - datasets.builder - Found cached dataset json (/home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092) +Loading Dataset info from /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092 +06/26/2025 23:48:27 - INFO - datasets.info - Loading Dataset info from /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092 +[INFO|configuration_utils.py:733] 2025-06-26 23:48:28,434 >> loading configuration file config.json from cache at /home/iitm_admin/.cache/huggingface/hub/models--meta-llama--Llama-3.1-8B-Instruct/snapshots/0e9e39f249a16976918f6564b8830bc894c89659/config.json +[INFO|configuration_utils.py:821] 2025-06-26 23:48:28,435 >> Model config LlamaConfig { + "_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "additional_loss_layer": 16, + "alignment_matrices_path": null, + "apply_inverse": false, + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "contrastive_loss_temperature": 1.0, + "contrastive_loss_weight": 1.0, + "contrastive_pooling_type": "mean", + "distance_function": "cosine", + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "inject_Ws": false, + "intermediate_size": 14336, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 32, + "num_hidden_layers": 32, + "num_key_value_heads": 8, + "only_train_contrastive": false, + "only_train_language_modeling": true, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.44.0.dev0", + "unidirectional_contrastive_loss": false, + "use_cache": true, + "vocab_size": 128256 +} + +[INFO|tokenization_utils_base.py:2269] 2025-06-26 23:48:28,661 >> loading file tokenizer.json from cache at /home/iitm_admin/.cache/huggingface/hub/models--meta-llama--Llama-3.1-8B-Instruct/snapshots/0e9e39f249a16976918f6564b8830bc894c89659/tokenizer.json +[INFO|tokenization_utils_base.py:2269] 2025-06-26 23:48:28,661 >> loading file added_tokens.json from cache at None +[INFO|tokenization_utils_base.py:2269] 2025-06-26 23:48:28,661 >> loading file special_tokens_map.json from cache at /home/iitm_admin/.cache/huggingface/hub/models--meta-llama--Llama-3.1-8B-Instruct/snapshots/0e9e39f249a16976918f6564b8830bc894c89659/special_tokens_map.json +[INFO|tokenization_utils_base.py:2269] 2025-06-26 23:48:28,661 >> loading file tokenizer_config.json from cache at /home/iitm_admin/.cache/huggingface/hub/models--meta-llama--Llama-3.1-8B-Instruct/snapshots/0e9e39f249a16976918f6564b8830bc894c89659/tokenizer_config.json +[INFO|tokenization_utils_base.py:2513] 2025-06-26 23:48:28,919 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. +06/26/2025 23:48:28 - INFO - __main__ - Tokenizer is fast: True +[INFO|modeling_utils.py:3667] 2025-06-26 23:48:28,923 >> loading weights file model.safetensors from cache at /home/iitm_admin/.cache/huggingface/hub/models--meta-llama--Llama-3.1-8B-Instruct/snapshots/0e9e39f249a16976918f6564b8830bc894c89659/model.safetensors.index.json +[INFO|modeling_utils.py:1591] 2025-06-26 23:48:28,924 >> Instantiating LlamaForCausalLM model under default dtype torch.bfloat16. +[WARNING|logging.py:328] 2025-06-26 23:48:28,927 >> You are attempting to use Flash Attention 2.0 with a model not initialized on GPU. Make sure to move the model to GPU after initializing it on CPU with `model.to('cuda')`. +[INFO|configuration_utils.py:1038] 2025-06-26 23:48:28,928 >> Generate config GenerationConfig { + "bos_token_id": 128000, + "eos_token_id": [ + 128001, + 128008, + 128009 + ] +} + +[WARNING|logging.py:328] 2025-06-26 23:48:28,930 >> You are attempting to use Flash Attention 2.0 with a model not initialized on GPU. Make sure to move the model to GPU after initializing it on CPU with `model.to('cuda')`. + Loading checkpoint shards: 0%| | 0/4 [00:00> You are attempting to use Flash Attention 2.0 with a model not initialized on GPU. Make sure to move the model to GPU after initializing it on CPU with `model.to('cuda')`. + Loading checkpoint shards: 50%|█████ | 2/4 [00:00<00:00, 4.72it/s] Loading checkpoint shards: 50%|█████ | 2/4 [00:00<00:00, 4.71it/s] Loading checkpoint shards: 0%| | 0/4 [00:00> All model checkpoint weights were used when initializing LlamaForCausalLM. + +[INFO|modeling_utils.py:4507] 2025-06-26 23:48:29,827 >> All the weights of LlamaForCausalLM were initialized from the model checkpoint at meta-llama/Llama-3.1-8B-Instruct. +If your task is similar to the task the model of the checkpoint was trained on, you can already use LlamaForCausalLM for predictions without further training. +[WARNING|logging.py:328] 2025-06-26 23:48:29,863 >> You are attempting to use Flash Attention 2.0 with a model not initialized on GPU. Make sure to move the model to GPU after initializing it on CPU with `model.to('cuda')`. + Loading checkpoint shards: 50%|█████ | 2/4 [00:00<00:00, 4.70it/s] Loading checkpoint shards: 0%| | 0/4 [00:00> loading configuration file generation_config.json from cache at /home/iitm_admin/.cache/huggingface/hub/models--meta-llama--Llama-3.1-8B-Instruct/snapshots/0e9e39f249a16976918f6564b8830bc894c89659/generation_config.json +[INFO|configuration_utils.py:1038] 2025-06-26 23:48:30,060 >> Generate config GenerationConfig { + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "temperature": 0.6, + "top_p": 0.9 +} + +adding special tokens... +06/26/2025 23:48:30 - INFO - __main__ - ================ pad, eos, bos, unk, padding ================ +06/26/2025 23:48:30 - INFO - __main__ - <|eot_id|>, 128009 +06/26/2025 23:48:30 - INFO - __main__ - <|eot_id|>, 128009 +06/26/2025 23:48:30 - INFO - __main__ - <|begin_of_text|>, 128000 +06/26/2025 23:48:30 - INFO - __main__ - <|reserved_special_token_0|>, 128002 +06/26/2025 23:48:30 - INFO - __main__ - right +06/26/2025 23:48:30 - INFO - __main__ - lora_r : 8 +06/26/2025 23:48:30 - INFO - __main__ - lora_alpha : 16 +06/26/2025 23:48:30 - INFO - __main__ - lora_dropout : 0.1 +06/26/2025 23:48:30 - INFO - __main__ - lora_target_modules : ['q_proj', 'k_proj', 'v_proj', 'o_proj', 'gate_proj', 'up_proj', 'down_proj'] +06/26/2025 23:48:30 - INFO - __main__ - LoRA configs: LoraConfig(task_type='CAUSAL_LM', peft_type=, auto_mapping=None, base_model_name_or_path=None, revision=None, inference_mode=False, r=8, target_modules={'v_proj', 'down_proj', 'k_proj', 'up_proj', 'q_proj', 'o_proj', 'gate_proj'}, exclude_modules=None, lora_alpha=16, lora_dropout=0.1, fan_in_fan_out=False, bias='none', use_rslora=False, modules_to_save=None, init_lora_weights=True, layers_to_transform=None, layers_pattern=None, rank_pattern={}, alpha_pattern={}, megatron_config=None, megatron_core='megatron.core', trainable_token_indices=None, loftq_config={}, eva_config=None, corda_config=None, use_dora=False, layer_replication=None, runtime_config=LoraRuntimeConfig(ephemeral_gpu_offload=False), lora_bias=False) + Loading checkpoint shards: 100%|██████████| 4/4 [00:00<00:00, 6.49it/s] Loading checkpoint shards: 100%|██████████| 4/4 [00:00<00:00, 5.79it/s] + Loading checkpoint shards: 25%|██▌ | 1/4 [00:00<00:00, 3.82it/s]trainable params: 20,971,520 || all params: 8,051,232,768 || trainable%: 0.2605 +PeftModelForCausalLM( + (base_model): LoraModel( + (model): LlamaForCausalLM( + (model): LlamaModel( + (embed_tokens): Embedding(128256, 4096) + (layers): ModuleList( + (0-31): 32 x LlamaDecoderLayer( + (self_attn): LlamaFlashAttention2( + (q_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (k_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (v_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (o_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (rotary_emb): LlamaRotaryEmbedding() + ) + (mlp): LlamaMLP( + (gate_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (up_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (down_proj): lora.Linear( + (base_layer): Linear(in_features=14336, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=14336, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (act_fn): SiLU() + ) + (input_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + (post_attention_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + ) + ) + (norm): LlamaRMSNorm((4096,), eps=1e-05) + (rotary_emb): LlamaRotaryEmbedding() + ) + (lm_head): Linear(in_features=4096, out_features=128256, bias=False) + ) + ) +) +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +[rank1]:[W626 23:48:30.009600268 ProcessGroupNCCL.cpp:4718] [PG ID 0 PG GUID 0 Rank 1] using GPU 1 as device used by this process is currently unknown. This can potentially cause a hang if this rank to GPU mapping is incorrect. You can pecify device_id in init_process_group() to force use of a particular device. + Loading checkpoint shards: 50%|█████ | 2/4 [00:00<00:00, 4.96it/s]trainable params: 20,971,520 || all params: 8,051,232,768 || trainable%: 0.2605 +PeftModelForCausalLM( + (base_model): LoraModel( + (model): LlamaForCausalLM( + (model): LlamaModel( + (embed_tokens): Embedding(128256, 4096) + (layers): ModuleList( + (0-31): 32 x LlamaDecoderLayer( + (self_attn): LlamaFlashAttention2( + (q_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (k_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (v_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (o_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (rotary_emb): LlamaRotaryEmbedding() + ) + (mlp): LlamaMLP( + (gate_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (up_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (down_proj): lora.Linear( + (base_layer): Linear(in_features=14336, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=14336, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (act_fn): SiLU() + ) + (input_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + (post_attention_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + ) + ) + (norm): LlamaRMSNorm((4096,), eps=1e-05) + (rotary_emb): LlamaRotaryEmbedding() + ) + (lm_head): Linear(in_features=4096, out_features=128256, bias=False) + ) + ) +) +06/26/2025 23:48:30 - INFO - __main__ - block size: 2048 +adding special tokens... +Loading cached processed dataset at /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092/cache-d04c1a71da869374.arrow +06/26/2025 23:48:30 - INFO - datasets.arrow_dataset - Loading cached processed dataset at /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092/cache-d04c1a71da869374.arrow + Loading checkpoint shards: 75%|███████▌ | 3/4 [00:00<00:00, 5.58it/s]Loading cached processed dataset at /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092/cache-930874df0ba118e5.arrow +06/26/2025 23:48:30 - INFO - datasets.arrow_dataset - Loading cached processed dataset at /home/iitm_admin/.cache/huggingface/datasets/json/default-10eaf7c5c1c6f11a/0.0.0/f4e89e8750d5d5ffbef2c078bf0ddfedef29dc2faff52a6255cf513c05eb1092/cache-930874df0ba118e5.arrow +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once + Loading checkpoint shards: 100%|██████████| 4/4 [00:00<00:00, 6.03it/s] Loading checkpoint shards: 100%|██��███████| 4/4 [00:00<00:00, 5.56it/s] +trainable params: 20,971,520 || all params: 8,051,232,768 || trainable%: 0.2605 +PeftModelForCausalLM( + (base_model): LoraModel( + (model): LlamaForCausalLM( + (model): LlamaModel( + (embed_tokens): Embedding(128256, 4096) + (layers): ModuleList( + (0-31): 32 x LlamaDecoderLayer( + (self_attn): LlamaFlashAttention2( + (q_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (k_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (v_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (o_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (rotary_emb): LlamaRotaryEmbedding() + ) + (mlp): LlamaMLP( + (gate_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (up_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (down_proj): lora.Linear( + (base_layer): Linear(in_features=14336, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=14336, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (act_fn): SiLU() + ) + (input_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + (post_attention_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + ) + ) + (norm): LlamaRMSNorm((4096,), eps=1e-05) + (rotary_emb): LlamaRotaryEmbedding() + ) + (lm_head): Linear(in_features=4096, out_features=128256, bias=False) + ) + ) +) +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +[rank3]:[W626 23:48:30.467276627 ProcessGroupNCCL.cpp:4718] [PG ID 0 PG GUID 0 Rank 3] using GPU 3 as device used by this process is currently unknown. This can potentially cause a hang if this rank to GPU mapping is incorrect. You can pecify device_id in init_process_group() to force use of a particular device. +[rank0]:[W626 23:48:30.494903331 ProcessGroupNCCL.cpp:4718] [PG ID 0 PG GUID 0 Rank 0] using GPU 0 as device used by this process is currently unknown. This can potentially cause a hang if this rank to GPU mapping is incorrect. You can pecify device_id in init_process_group() to force use of a particular device. +adding special tokens... +trainable params: 20,971,520 || all params: 8,051,232,768 || trainable%: 0.2605 +PeftModelForCausalLM( + (base_model): LoraModel( + (model): LlamaForCausalLM( + (model): LlamaModel( + (embed_tokens): Embedding(128256, 4096) + (layers): ModuleList( + (0-31): 32 x LlamaDecoderLayer( + (self_attn): LlamaFlashAttention2( + (q_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (k_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (v_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=1024, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=1024, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (o_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (rotary_emb): LlamaRotaryEmbedding() + ) + (mlp): LlamaMLP( + (gate_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (up_proj): lora.Linear( + (base_layer): Linear(in_features=4096, out_features=14336, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=4096, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=14336, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (down_proj): lora.Linear( + (base_layer): Linear(in_features=14336, out_features=4096, bias=False) + (lora_dropout): ModuleDict( + (default): Dropout(p=0.1, inplace=False) + ) + (lora_A): ModuleDict( + (default): Linear(in_features=14336, out_features=8, bias=False) + ) + (lora_B): ModuleDict( + (default): Linear(in_features=8, out_features=4096, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (act_fn): SiLU() + ) + (input_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + (post_attention_layernorm): LlamaRMSNorm((4096,), eps=1e-05) + ) + ) + (norm): LlamaRMSNorm((4096,), eps=1e-05) + (rotary_emb): LlamaRotaryEmbedding() + ) + (lm_head): Linear(in_features=4096, out_features=128256, bias=False) + ) + ) +) +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +[rank2]:[W626 23:48:31.966426045 ProcessGroupNCCL.cpp:4718] [PG ID 0 PG GUID 0 Rank 2] using GPU 2 as device used by this process is currently unknown. This can potentially cause a hang if this rank to GPU mapping is incorrect. You can pecify device_id in init_process_group() to force use of a particular device. +06/26/2025 23:48:32 - INFO - __main__ - xxx: Showcase the tokenized training samples. +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 28{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 280, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]}4, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -4, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} + +100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100{'input_ids': [128000, 128006, 9125, 128007, 271, 11372, 228, 11372, 103, 87648, 62456, 36278, 237, 11372, 243, 11372, 250, 87648, 36278, 116, 11372, 117, 50228, 107, 11372, 120, 11372, 243, 36278, 245, 11372, 96, 81278, 97, 36278, 114, 81278, 243, 53906, 115, 11372, 243, 60008, 73358, 36278, 255, 28025, 224, 11372, 106, 81278, 243, 50228, 107, 11372, 120, 36278, 228, 11372, 249, 60008, 87648, 100278, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 250, 87648, 53906, 107, 11, 36278, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 248, 81278, 101, 53906, 97, 42412, 36278, 243, 73358, 28025, 223, 87648, 36278, 237, 11372, 105, 11372, 224, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 116, 11372, 106, 50228, 100, 50228, 101, 11372, 253, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 36278, 241, 36278, 116, 11372, 224, 11372, 110, 50228, 103, 11372, 106, 28025, 224, 11372, 110, 11372, 243, 11372, 255, 50228, 105, 60008, 36278, 231, 11372, 103, 11372, 116, 53906, 98, 50228, 103, 87648, 36278, 243, 73358, 28025, 223, 87648, 100278, 36278, 228, 11372, 103, 87648, 50228, 108, 36278, 231, 11372, 97, 53906, 97, 73358, 60008, 73358, 36278, 114, 28025, 223, 73358, 28025, 223, 11372, 97, 60008, 330, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 3047, 36278, 110, 81278, 244, 28025, 223, 87648, 11, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 245, 11372, 96, 87648, 50228, 245, 28025, 223, 11372, 110, 62456, 36278, 116, 53906, 103, 11372, 115, 53906, 253, 11372, 255, 50228, 105, 60008, 36278, 99, 60008, 11372, 244, 50228, 101, 36278, 237, 11372, 105, 11372, 224, 36278, 114, 60008, 11372, 115, 60008, 36278, 110, 81278, 244, 28025, 223, 87648, 330, 11372, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 28025, 233, 1630, 100278, 498, 36278, 107, 60008, 11372, 244, 50228, 101, 60008, 1630, 36278, 117, 11372, 110, 28025, 233, 36278, 248, 28025, 224, 11372, 94, 11372, 120, 42412, 87648, 53906, 97, 36278, 104, 11372, 110, 50228, 104, 11372, 110, 100278, 36278, 107, 11372, 98, 50228, 108, 53906, 98, 11372, 97, 42412, 36278, 237, 11372, 105, 11372, 224, 36278, 105, 81278, 116, 53906, 97, 50228, 108, 81278, 97, 36278, 105, 53906, 107, 50228, 244, 53906, 107, 50228, 108, 36278, 103, 53906, 108, 11372, 97, 62456, 36278, 107, 11372, 97, 53906, 101, 11372, 114, 28025, 222, 11372, 110, 36278, 98, 50228, 243, 28025, 223, 87648, 100278, 128009, 198, 128006, 882, 128007, 271, 11372, 103, 53906, 108, 11372, 114, 53906, 101, 25, 220, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 400, 17, 11, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 400, 18, 11, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 60008, 73358, 36278, 99, 50228, 106, 400, 87, 36278, 237, 11372, 105, 11372, 224, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 103, 53906, 108, 11372, 97, 81278, 253, 62456, 73358, 36278, 99, 50228, 106, 400, 16, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 101, 81278, 250, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 237, 11372, 105, 11372, 224, 36278, 243, 81278, 249, 28025, 223, 36278, 105, 87648, 53906, 100, 28025, 223, 11372, 99, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 243, 81278, 249, 28025, 223, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 100278, 36278, 97, 81278, 101, 62456, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 42412, 11, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 11372, 116, 36278, 237, 11372, 105, 11372, 224, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 36278, 227, 73358, 53906, 94, 50228, 108, 36278, 243, 73358, 60008, 87648, 100278, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 60008, 73358, 36278, 250, 87648, 53906, 107, 36278, 116, 60008, 36278, 243, 11372, 97, 11372, 253, 42412, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 36278, 243, 73358, 60008, 30, 36278, 228, 11372, 106, 73358, 42412, 36278, 107, 11372, 99, 62456, 36278, 231, 11372, 103, 73358, 60008, 73358, 36278, 103, 53906, 108, 11372, 114, 53906, 101, 60008, 73358, 36278, 231, 11372, 97, 53906, 97, 73358, 220, 18, 36278, 250, 50228, 101, 62456, 36278, 97, 50228, 117, 11372, 110, 60008, 36278, 227, 11372, 250, 50228, 101, 42412, 36278, 248, 11372, 110, 11372, 243, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 243, 11372, 97, 30, 128009, 198, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -10, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +0, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 128006, 78191, 128007, 271, 11372, 100, 50228, 103, 60008, 36278, 100, 50228, 103, 60008, 36278, 231, 11372, 97, 53906, 97, 73358, 25, 36278, 97, 81278, 101, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 103, 28025, 223, 11372, 248, 81278, 101, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 18, 353, 400, 17, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 228, 11372, 229, 11372, 116, 11372, 94, 36278, 248, 50228, 107, 11372, 120, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 18, 284, 400, 21, 100278, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 243, 53906, 107, 50228, 104, 60008, 36278, 110, 53906, 107, 50228, 253, 60008, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 87, 284, 400, 17, 87, 28025, 115, 36278, 99, 28025, 223, 11372, 253, 62456, 36278, 237, 11372, 116, 11372, 103, 53906, 108, 60008, 11372, 116, 28025, 233, 73358, 36278, 99, 50228, 106, 220, 17, 353, 400, 16, 284, 400, 17, 100278, 36278, 116, 11372, 106, 11372, 116, 53906, 97, 36278, 103, 50228, 101, 28025, 222, 11372, 107, 11372, 120, 60008, 73358, 36278, 106, 28025, 233, 11372, 253, 36278, 106, 28025, 224, 11372, 110, 53906, 107, 36278, 117, 11372, 110, 400, 21, 489, 400, 21, 489, 400, 17, 87, 489, 400, 17, 284, 400, 975, 489, 400, 17, 87, 100278, 36278, 116, 53906, 107, 50228, 101, 53906, 94, 62456, 36278, 105, 81278, 114, 36278, 94, 11372, 110, 50228, 108, 60008, 73358, 36278, 105, 81278, 110, 36278, 99, 81278, 107, 11372, 120, 60008, 36278, 227, 73358, 53906, 98, 36278, 103, 53906, 108, 11372, 99, 50228, 101, 36278, 243, 73358, 60008, 11, 36278, 97, 50228, 229, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 104, 60008, 73358, 11372, 97, 36278, 103, 50228, 107, 11372, 120, 36278, 97, 42412, 36278, 117, 11372, 110, 400, 508, 482, 1746, 975, 489, 400, 17, 87, 8, 284, 400, 508, 482, 400, 975, 482, 400, 17, 87, 284, 400, 21, 482, 400, 17, 87, 28025, 115, 36278, 228, 11372, 106, 50228, 99, 60008, 73358, 36278, 99, 60008, 11372, 241, 11372, 107, 11372, 120, 42412, 36278, 117, 11372, 107, 11372, 120, 60008, 11372, 249, 60008, 36278, 107, 60008, 36278, 116, 60008, 36278, 107, 60008, 36278, 103, 73358, 81278, 105, 73358, 53906, 97, 87648, 11372, 253, 62456, 36278, 103, 50228, 105, 60008, 36278, 97, 42412, 36278, 117, 11372, 110, 220, 18, 36278, 94, 11372, 110, 50228, 108, 11, 36278, 97, 50228, 229, 36278, 228, 11372, 106, 73358, 42412, 36278, 110, 81278, 244, 11372, 97, 60008, 36278, 103, 50228, 108, 62456, 25, 400, 21, 482, 400, 17, 87, 284, 400, 18, 28025, 115, 36278, 116, 73358, 11372, 110, 28025, 222, 11372, 243, 73358, 11372, 96, 36278, 243, 73358, 60008, 11, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 400, 17, 87, 284, 400, 18, 100278, 36278, 231, 11372, 255, 11372, 107, 11372, 120, 36278, 103, 11372, 243, 53906, 115, 11372, 243, 60008, 400, 17, 36278, 99, 53906, 105, 50228, 108, 42412, 36278, 255, 50228, 245, 36278, 243, 73358, 11372, 110, 60008, 36278, 228, 11372, 106, 73358, 42412, 36278, 103, 50228, 229, 25, 865, 284, 400, 16, 13, 1135, 100278, 865, 36278, 237, 73358, 36278, 106, 50228, 101, 36278, 117, 11372, 110, 400, 16, 13, 1135, 100278, 36278, 231, 11372, 97, 53906, 97, 73358, 36278, 117, 11372, 110, 220, 16, 13, 20, 100278, 128009]} +/home/iitm_admin/llmteam/mid-align/src/transformers/utils/import_utils.py:560: FutureWarning: `is_torch_tpu_available` is deprecated and will be removed in 4.41.0. Please use the `is_torch_xla_available` instead. + warnings.warn( +/home/iitm_admin/llmteam/mid-align/src/transformers/utils/import_utils.py:560: FutureWarning: `is_torch_tpu_available` is deprecated and will be removed in 4.41.0. Please use the `is_torch_xla_available` instead. + warnings.warn( +/home/iitm_admin/llmteam/mid-align/src/transformers/deepspeed.py:24: FutureWarning: transformers.deepspeed module is deprecated and will be removed in a future version. Please import deepspeed modules directly from transformers.integrations + warnings.warn( +/home/iitm_admin/llmteam/mid-align/src/transformers/deepspeed.py:24: FutureWarning: transformers.deepspeed module is deprecated and will be removed in a future version. Please import deepspeed modules directly from transformers.integrations + warnings.warn( +/home/iitm_admin/llmteam/mid-align/src/transformers/utils/import_utils.py:560: FutureWarning: `is_torch_tpu_available` is deprecated and will be removed in 4.41.0. Please use the `is_torch_xla_available` instead. + warnings.warn( +/home/iitm_admin/llmteam/mid-align/src/transformers/deepspeed.py:24: FutureWarning: transformers.deepspeed module is deprecated and will be removed in a future version. Please import deepspeed modules directly from transformers.integrations + warnings.warn( +/home/iitm_admin/llmteam/mid-align/src/transformers/utils/import_utils.py:560: FutureWarning: `is_torch_tpu_available` is deprecated and will be removed in 4.41.0. Please use the `is_torch_xla_available` instead. + warnings.warn( +/home/iitm_admin/llmteam/mid-align/src/transformers/deepspeed.py:24: FutureWarning: transformers.deepspeed module is deprecated and will be removed in a future version. Please import deepspeed modules directly from transformers.integrations + warnings.warn( +[INFO|trainer.py:658] 2025-06-26 23:48:34,726 >> Using auto half precision backend +[2025-06-26 23:48:34,996] [INFO] [logging.py:107:log_dist] [Rank 0] DeepSpeed info: version=0.17.1, git-hash=unknown, git-branch=unknown +[2025-06-26 23:48:34,996] [INFO] [config.py:655:__init__] Config mesh_device None world_size = 4 +[2025-06-26 23:48:39,308] [INFO] [engine.py:1325:_configure_distributed_model] ********** distributed groups summary ********** + self.dp_world_size=4 + self.mp_world_size=1 + self.seq_dp_world_size=4 + self.sequence_parallel_size=1 +*********************************************** +[2025-06-26 23:48:39,902] [INFO] [logging.py:107:log_dist] [Rank 0] DeepSpeed Flops Profiler Enabled: False +Installed CUDA version 12.0 does not match the version torch was compiled with 12.6 but since the APIs are compatible, accepting this combination +Using /home/iitm_admin/.cache/torch_extensions/py39_cu126 as PyTorch extensions root... +Installed CUDA version 12.0 does not match the version torch was compiled with 12.6 but since the APIs are compatible, accepting this combination +Using /home/iitm_admin/.cache/torch_extensions/py39_cu126 as PyTorch extensions root... +Detected CUDA files, patching ldflags +Emitting ninja build file /home/iitm_admin/.cache/torch_extensions/py39_cu126/cpu_adam/build.ninja... +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation. +If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST']. + warnings.warn( +Building extension module cpu_adam... +Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) +Installed CUDA version 12.0 does not match the version torch was compiled with 12.6 but since the APIs are compatible, accepting this combination +Using /home/iitm_admin/.cache/torch_extensions/py39_cu126 as PyTorch extensions root... +Installed CUDA version 12.0 does not match the version torch was compiled with 12.6 but since the APIs are compatible, accepting this combination +Using /home/iitm_admin/.cache/torch_extensions/py39_cu126 as PyTorch extensions root... +ninja: no work to do. +Loading extension module cpu_adam... +Time to load cpu_adam op: 2.7387800216674805 seconds +Adam Optimizer #0 is created with AVX512 arithmetic capability. +Config: alpha=0.000500, betas=(0.900000, 0.999000), weight_decay=0.000000, adam_w=1 +[2025-06-26 23:48:44,061] [INFO] [logging.py:107:log_dist] [Rank 0] Using DeepSpeed Optimizer param name adam as basic optimizer +[2025-06-26 23:48:44,061] [INFO] [logging.py:107:log_dist] [Rank 0] Removing param_group that has no 'params' in the basic Optimizer +[2025-06-26 23:48:44,109] [INFO] [logging.py:107:log_dist] [Rank 0] DeepSpeed Basic Optimizer = DeepSpeedCPUAdam +[2025-06-26 23:48:44,110] [INFO] [utils.py:59:is_zero_supported_optimizer] Checking ZeRO support for optimizer=DeepSpeedCPUAdam type= +[2025-06-26 23:48:44,110] [INFO] [logging.py:107:log_dist] [Rank 0] Creating torch.bfloat16 ZeRO stage 1 optimizer +[2025-06-26 23:48:44,110] [INFO] [stage_1_and_2.py:151:__init__] Reduce bucket size 200000000 +[2025-06-26 23:48:44,110] [INFO] [stage_1_and_2.py:152:__init__] Allgather bucket size 200000000 +[2025-06-26 23:48:44,110] [INFO] [stage_1_and_2.py:153:__init__] CPU Offload: True +[2025-06-26 23:48:44,110] [INFO] [stage_1_and_2.py:154:__init__] Round robin gradient partitioning: False +Loading extension module cpu_adam... +Loading extension module cpu_adam... +Time to load cpu_adam op: 2.7846410274505615 seconds +Time to load cpu_adam op: 2.8138020038604736 seconds +Loading extension module cpu_adam... +Time to load cpu_adam op: 2.7922706604003906 seconds +[2025-06-26 23:48:44,515] [INFO] [utils.py:781:see_memory_usage] Before initializing optimizer states +[2025-06-26 23:48:44,515] [INFO] [utils.py:782:see_memory_usage] MA 15.0 GB Max_MA 15.0 GB CA 15.16 GB Max_CA 15 GB +[2025-06-26 23:48:44,516] [INFO] [utils.py:789:see_memory_usage] CPU Virtual Memory: used = 77.31 GB, percent = 3.8% +[2025-06-26 23:48:44,744] [INFO] [utils.py:781:see_memory_usage] After initializing optimizer states +[2025-06-26 23:48:44,745] [INFO] [utils.py:782:see_memory_usage] MA 15.0 GB Max_MA 15.0 GB CA 15.16 GB Max_CA 15 GB +[2025-06-26 23:48:44,745] [INFO] [utils.py:789:see_memory_usage] CPU Virtual Memory: used = 77.44 GB, percent = 3.8% +[2025-06-26 23:48:44,745] [INFO] [stage_1_and_2.py:573:__init__] optimizer state initialized +[2025-06-26 23:48:44,906] [INFO] [utils.py:781:see_memory_usage] After initializing ZeRO optimizer +[2025-06-26 23:48:44,907] [INFO] [utils.py:782:see_memory_usage] MA 15.0 GB Max_MA 15.0 GB CA 15.16 GB Max_CA 15 GB +[2025-06-26 23:48:44,907] [INFO] [utils.py:789:see_memory_usage] CPU Virtual Memory: used = 77.52 GB, percent = 3.8% +[2025-06-26 23:48:44,910] [INFO] [logging.py:107:log_dist] [Rank 0] DeepSpeed Final Optimizer = DeepSpeedZeroOptimizer +[2025-06-26 23:48:44,910] [INFO] [logging.py:107:log_dist] [Rank 0] DeepSpeed using client callable to create LR scheduler +[2025-06-26 23:48:44,910] [INFO] [logging.py:107:log_dist] [Rank 0] DeepSpeed LR Scheduler = +[2025-06-26 23:48:44,910] [INFO] [logging.py:107:log_dist] [Rank 0] step=0, skipped=0, lr=[0.0], mom=[[0.9, 0.999]] +[2025-06-26 23:48:44,916] [INFO] [logging.py:107:log_dist] [Rank 0] [TorchCheckpointEngine] Initialized with serialization = True +[2025-06-26 23:48:44,916] [INFO] [config.py:921:print] DeepSpeedEngine configuration: +[2025-06-26 23:48:44,916] [INFO] [config.py:925:print] activation_checkpointing_config { + "partition_activations": false, + "contiguous_memory_optimization": false, + "cpu_checkpointing": false, + "number_checkpoints": null, + "synchronize_checkpoint_boundary": false, + "profile": false +} +[2025-06-26 23:48:44,916] [INFO] [config.py:925:print] aio_config ................... {'block_size': 1048576, 'queue_depth': 8, 'intra_op_parallelism': 1, 'single_submit': False, 'overlap_events': True, 'use_gds': False} +[2025-06-26 23:48:44,916] [INFO] [config.py:925:print] amp_enabled .................. False +[2025-06-26 23:48:44,916] [INFO] [config.py:925:print] amp_params ................... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] autotuning_config ............ { + "enabled": false, + "start_step": null, + "end_step": null, + "metric_path": null, + "arg_mappings": null, + "metric": "throughput", + "model_info": null, + "results_dir": "autotuning_results", + "exps_dir": "autotuning_exps", + "overwrite": true, + "fast": true, + "start_profile_step": 3, + "end_profile_step": 5, + "tuner_type": "gridsearch", + "tuner_early_stopping": 5, + "tuner_num_trials": 50, + "model_info_path": null, + "mp_size": 1, + "max_train_batch_size": null, + "min_train_batch_size": 1, + "max_train_micro_batch_size_per_gpu": 1.024000e+03, + "min_train_micro_batch_size_per_gpu": 1, + "num_tuning_micro_batch_sizes": 3 +} +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] bfloat16_config .............. enabled=True immediate_grad_update=False check_grad_overflow=False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] checkpoint_config ............ {'tag_validation': 'WARN', 'checkpoint_serialization': True, 'writer': None} +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] checkpoint_parallel_write_pipeline False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] checkpoint_tag_validation_enabled True +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] checkpoint_tag_validation_fail False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] comms_config ................. +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] communication_data_type ...... None +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] compile_config ............... deepcompile=False free_activation=False offload_activation=False offload_opt_states=False double_buffer=True symmetric_memory=False debug_log=False offload_parameters=False sync_before_reduce=False sync_after_reduce=False sync_before_allgather=False sync_after_allgather=False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] compression_config ........... {'weight_quantization': {'shared_parameters': {'enabled': False, 'quantizer_kernel': False, 'schedule_offset': 0, 'quantize_groups': 1, 'quantize_verbose': False, 'quantization_type': 'symmetric', 'quantize_weight_in_forward': False, 'rounding': 'nearest', 'fp16_mixed_quantize': False, 'quantize_change_ratio': 0.001}, 'different_groups': {}}, 'activation_quantization': {'shared_parameters': {'enabled': False, 'quantization_type': 'symmetric', 'range_calibration': 'dynamic', 'schedule_offset': 1000}, 'different_groups': {}}, 'sparse_pruning': {'shared_parameters': {'enabled': False, 'method': 'l1', 'schedule_offset': 1000}, 'different_groups': {}}, 'row_pruning': {'shared_parameters': {'enabled': False, 'method': 'l1', 'schedule_offset': 1000}, 'different_groups': {}}, 'head_pruning': {'shared_parameters': {'enabled': False, 'method': 'topk', 'schedule_offset': 1000}, 'different_groups': {}}, 'channel_pruning': {'shared_parameters': {'enabled': False, 'method': 'l1', 'schedule_offset': 1000}, 'different_groups': {}}, 'layer_reduction': {'enabled': False}} +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] curriculum_enabled_legacy .... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] curriculum_params_legacy ..... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] data_efficiency_config ....... {'enabled': False, 'seed': 1234, 'data_sampling': {'enabled': False, 'num_epochs': 1000, 'num_workers': 0, 'pin_memory': False, 'curriculum_learning': {'enabled': False}, 'dynamic_batching': {'enabled': False, 'lr_scaling_method': 'linear', 'min_batch_size': 1, 'max_batch_size': None, 'sequence_picking_order': 'dataloader', 'verbose': False}}, 'data_routing': {'enabled': False, 'random_ltd': {'enabled': False, 'layer_token_lr_schedule': {'enabled': False}}}} +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] data_efficiency_enabled ...... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] dataloader_drop_last ......... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] disable_allgather ............ False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] dump_state ................... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_enabled ........... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_gas_boundary_resolution 1 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_layer_name ........ bert.encoder.layer +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_layer_num ......... 0 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_max_iter .......... 100 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_stability ......... 1e-06 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_tol ............... 0.01 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] eigenvalue_verbose ........... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] elasticity_enabled ........... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] float16_config ............... enabled=False auto_cast=False loss_scale=0.0 initial_scale_power=16 loss_scale_window=1000 hysteresis=2 consecutive_hysteresis=False min_loss_scale=1 fp16_master_weights_and_grads=False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] flops_profiler_config ........ { + "enabled": false, + "recompute_fwd_factor": 0.0, + "profile_step": 1, + "module_depth": -1, + "top_modules": 1, + "detailed": true, + "output_file": null +} +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] global_rank .................. 0 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] grad_accum_dtype ............. None +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] gradient_accumulation_steps .. 1 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] gradient_clipping ............ 1.0 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] gradient_predivide_factor .... 1.0 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] graph_harvesting ............. False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] hybrid_engine ................ enabled=False max_out_tokens=512 inference_tp_size=1 release_inference_cache=False pin_parameters=True tp_gather_partition_size=8 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] load_universal_checkpoint .... False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] memory_breakdown ............. False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] mics_hierarchial_params_gather False +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] mics_shard_size .............. -1 +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] monitor_config ............... tensorboard=TensorBoardConfig(enabled=False, output_path='', job_name='DeepSpeedJobName') comet=CometConfig(enabled=False, samples_log_interval=100, project=None, workspace=None, api_key=None, experiment_name=None, experiment_key=None, online=None, mode=None) wandb=WandbConfig(enabled=False, group=None, team=None, project='deepspeed') csv_monitor=CSVConfig(enabled=False, output_path='', job_name='DeepSpeedJobName') +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] nebula_config ................ { + "enabled": false, + "persistent_storage_path": null, + "persistent_time_interval": 100, + "num_of_version_in_retention": 2, + "enable_nebula_load": true, + "load_path": null +} +[2025-06-26 23:48:44,917] [INFO] [config.py:925:print] optimizer_legacy_fusion ...... False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] optimizer_name ............... adam +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] optimizer_params ............. {'lr': 0.0005, 'betas': [0.9, 0.999], 'eps': 1e-08, 'weight_decay': 0.0} +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] pipeline ..................... {'stages': 'auto', 'partition': 'best', 'seed_layers': False, 'activation_checkpoint_interval': 0, 'pipe_partitioned': True, 'grad_partitioned': True} +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] pld_enabled .................. False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] pld_params ................... False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] prescale_gradients ........... False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] scheduler_name ............... None +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] scheduler_params ............. None +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] seq_parallel_communication_data_type torch.float32 +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] sparse_attention ............. None +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] sparse_gradients_enabled ..... False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] steps_per_print .............. inf +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] tensor_parallel_config ....... dtype=torch.float16 autotp_size=0 tp_overlap_comm=False tensor_parallel=TPConfig(tp_size=1, tp_grain_size=1, mpu=None, tp_group=None) injection_policy_tuple=None keep_module_on_host=False replace_with_kernel_inject=False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] timers_config ................ enabled=True synchronized=True +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] train_batch_size ............. 100 +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] train_micro_batch_size_per_gpu 25 +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] use_data_before_expert_parallel_ False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] use_node_local_storage ....... False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] wall_clock_breakdown ......... False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] weight_quantization_config ... None +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] world_size ................... 4 +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] zero_allow_untested_optimizer False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] zero_config .................. stage=1 contiguous_gradients=True reduce_scatter=True reduce_bucket_size=200000000 use_multi_rank_bucket_allreduce=True allgather_partitions=True allgather_bucket_size=200000000 overlap_comm=True load_from_fp32_weights=True elastic_checkpoint=False offload_param=None offload_optimizer=DeepSpeedZeroOffloadOptimizerConfig(device='cpu', nvme_path=None, buffer_count=4, pin_memory=True, pipeline_read=False, pipeline_write=False, fast_init=False, ratio=1.0) sub_group_size=1000000000 cpu_offload_param=None cpu_offload_use_pin_memory=None cpu_offload=None prefetch_bucket_size=50000000 param_persistence_threshold=100000 model_persistence_threshold=9223372036854775807 max_live_parameters=1000000000 max_reuse_distance=1000000000 gather_16bit_weights_on_model_save=False module_granularity_threshold=0 use_all_reduce_for_fetch_params=False stage3_gather_fp16_weights_on_model_save=False ignore_unused_parameters=True legacy_stage1=False round_robin_gradients=False zero_hpz_partition_size=1 zero_quantized_weights=False zero_quantized_nontrainable_weights=False zero_quantized_gradients=False zeropp_loco_param=None mics_shard_size=-1 mics_hierarchical_params_gather=False memory_efficient_linear=True pipeline_loading_checkpoint=False override_module_apply=True log_trace_cache_warnings=False +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] zero_enabled ................. True +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] zero_force_ds_cpu_optimizer .. True +[2025-06-26 23:48:44,918] [INFO] [config.py:925:print] zero_optimization_stage ...... 1 +[2025-06-26 23:48:44,918] [INFO] [config.py:911:print_user_config] json = { + "optimizer": { + "type": "Adam", + "params": { + "lr": 0.0005, + "betas": [0.9, 0.999], + "eps": 1e-08, + "weight_decay": 0.0 + } + }, + "bf16": { + "enabled": true + }, + "fp16": { + "enabled": false, + "loss_scale": 0, + "loss_scale_window": 1000, + "initial_scale_power": 16, + "hysteresis": 2, + "min_loss_scale": 1 + }, + "zero_optimization": { + "stage": 1, + "offload_optimizer": { + "device": "cpu", + "pin_memory": true + }, + "allgather_partitions": true, + "allgather_bucket_size": 2.000000e+08, + "overlap_comm": true, + "reduce_scatter": true, + "reduce_bucket_size": 2.000000e+08, + "contiguous_gradients": true + }, + "gradient_accumulation_steps": 1, + "gradient_clipping": 1.0, + "steps_per_print": inf, + "train_batch_size": 100, + "train_micro_batch_size_per_gpu": 25, + "wall_clock_breakdown": false +} +[INFO|trainer.py:2145] 2025-06-26 23:48:44,920 >> ***** Running training ***** +[INFO|trainer.py:2146] 2025-06-26 23:48:44,920 >> Num examples = 15,000 +[INFO|trainer.py:2147] 2025-06-26 23:48:44,920 >> Num Epochs = 5 +[INFO|trainer.py:2148] 2025-06-26 23:48:44,920 >> Instantaneous batch size per device = 25 +[INFO|trainer.py:2151] 2025-06-26 23:48:44,920 >> Total train batch size (w. parallel, distributed & accumulation) = 100 +[INFO|trainer.py:2152] 2025-06-26 23:48:44,920 >> Gradient Accumulation steps = 1 +[INFO|trainer.py:2153] 2025-06-26 23:48:44,920 >> Total optimization steps = 750 +[INFO|trainer.py:2154] 2025-06-26 23:48:44,924 >> Number of trainable parameters = 20,971,520 +[INFO|integration_utils.py:807] 2025-06-26 23:48:44,927 >> Automatic Weights & Biases logging enabled, to disable set os.environ["WANDB_DISABLED"] = "true" +wandb: WARNING The `run_name` is currently set to the same value as `TrainingArguments.output_dir`. If this was not intended, please specify a different run name by setting the `TrainingArguments.run_name` parameter. +wandb: Currently logged in as: sidharthpulipaka (indic-encoder) to https://api.wandb.ai. Use `wandb login --relogin` to force relogin +wandb: Tracking run with wandb version 0.20.1 +wandb: Run data is saved locally in /home/iitm_admin/llmteam/mid-align/wandb/run-20250626_234845-s645lnzf +wandb: Run `wandb offline` to turn off syncing. +wandb: Syncing run ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/ +wandb: ⭐️ View project at https://wandb.ai/indic-encoder/midalign +wandb: 🚀 View run at https://wandb.ai/indic-encoder/midalign/runs/s645lnzf + 0%| | 0/750 [00:00> +***** Running Evaluation ***** +[INFO|trainer.py:3833] 2025-06-26 23:48:46,992 >> Num examples = 1000 +[INFO|trainer.py:3836] 2025-06-26 23:48:46,992 >> Batch size = 25 + + 0%| | 0/10 [00:00> +***** Running Evaluation ***** +[INFO|trainer.py:3833] 2025-06-27 00:09:28,330 >> Num examples = 1000 +[INFO|trainer.py:3836] 2025-06-27 00:09:28,330 >> Batch size = 25 + + 0%| | 0/10 [00:00> Saving model checkpoint to ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-200 +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d940c-6714583f6819fba559eba40a;6ea89e93-e0ba-4350-8087-8c66ff135abc) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +[INFO|tokenization_utils_base.py:2684] 2025-06-27 00:10:12,935 >> tokenizer config file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-200/tokenizer_config.json +[INFO|tokenization_utils_base.py:2693] 2025-06-27 00:10:12,935 >> Special tokens file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-200/special_tokens_map.json +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +[2025-06-27 00:10:17,953] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step200 is begin to save! +[2025-06-27 00:10:17,976] [INFO] [logging.py:107:log_dist] [Rank 0] Saving model checkpoint: ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-200/global_step200/mp_rank_00_model_states.pt +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9412-0525cb24664bc6a817a5199e;b69187b8-6c13-4836-9e8f-444b992560b1) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9412-79eea90b0080f36458fd28e4;c5330cdc-d9ad-4653-a60c-e9792beac3da) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9412-673b11620f41494e0798129d;1d3ca5be-5320-4237-a9ef-32d90bdc691b) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9412-47a335377afb587169c716ce;b53874b3-dd19-4254-8815-5a03032f26b0) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( + 27%|██▋ | 201/750 [21:37<3:12:07, 21.00s/it] {'loss': 0.2006, 'grad_norm': 0.09802627563476562, 'learning_rate': 0.0001691359369682545, 'epoch': 1.34} + 27%|██▋ | 201/750 [21:37<3:12:07, 21.00s/it] 27%|██▋ | 202/750 [21:43<2:30:12, 16.45s/it] {'loss': 0.187, 'grad_norm': 0.09598495066165924, 'learning_rate': 0.00016871676423714827, 'epoch': 1.35} + 27%|██▋ | 202/750 [21:43<2:30:12, 16.45s/it] 27%|██▋ | 203/750 [21:49<2:01:51, 13.37s/it] {'loss': 0.2075, 'grad_norm': 0.10226590186357498, 'learning_rate': 0.00016830069266853705, 'epoch': 1.35} + 27%|██▋ | 203/750 [21:49<2:01:51, 13.37s/it] 27%|██▋ | 204/750 [21:55<1:41:19, 11.13s/it] {'loss': 0.2012, 'grad_norm': 0.0975937768816948, 'learning_rate': 0.00016788768421121283, 'epoch': 1.36} + 27%|██▋ | 204/750 [21:55<1:41:19, 11.13s/it] 27%|██▋ | 205/750 [22:01<1:27:20, 9.62s/it] {'loss': 0.1976, 'grad_norm': 0.09934520721435547, 'learning_rate': 0.00016747770146441848, 'epoch': 1.37} + 27%|██▋ | 205/750 [22:01<1:27:20, 9.62s/it] 27%|██▋ | 206/750 [22:07<1:17:10, 8.51s/it] {'loss': 0.1941, 'grad_norm': 0.09936224669218063, 'learning_rate': 0.0001670707076636216, 'epoch': 1.37} + 27%|██▋ | 206/750 [22:07<1:17:10, 8.51s/it] 28%|██▊ | 207/750 [22:13<1:09:45, 7.71s/it] {'loss': 0.1997, 'grad_norm': 0.10343389958143234, 'learning_rate': 0.00016666666666666666, 'epoch': 1.38} + 28%|██▊ | 207/750 [22:13<1:09:45, 7.71s/it] 28%|██▊ | 208/750 [22:19<1:04:54, 7.18s/it] {'loss': 0.1948, 'grad_norm': 0.09763069450855255, 'learning_rate': 0.0001662655429402941, 'epoch': 1.39} + 28%|██▊ | 208/750 [22:19<1:04:54, 7.18s/it] 28%|██▊ | 209/750 [22:25<1:01:35, 6.83s/it] {'loss': 0.19, 'grad_norm': 0.08771580457687378, 'learning_rate': 0.00016586730154701388, 'epoch': 1.39} + 28%|██▊ | 209/750 [22:25<1:01:35, 6.83s/it] 28%|██▊ | 210/750 [22:31<59:06, 6.57s/it] {'loss': 0.1954, 'grad_norm': 0.10181500762701035, 'learning_rate': 0.00016547190813232432, 'epoch': 1.4} + 28%|██▊ | 210/750 [22:31<59:06, 6.57s/it] 28%|██▊ | 211/750 [22:37<57:33, 6.41s/it] {'loss': 0.1923, 'grad_norm': 0.0967947468161583, 'learning_rate': 0.00016507932891226336, 'epoch': 1.41} + 28%|██▊ | 211/750 [22:37<57:33, 6.41s/it] 28%|██▊ | 212/750 [22:43<55:56, 6.24s/it] {'loss': 0.1963, 'grad_norm': 0.09669908881187439, 'learning_rate': 0.00016468953066128386, 'epoch': 1.41} + 28%|██▊ | 212/750 [22:43<55:56, 6.24s/it] 28%|██▊ | 213/750 [22:49<55:14, 6.17s/it] {'loss': 0.2168, 'grad_norm': 0.10868806391954422, 'learning_rate': 0.00016430248070044244, 'epoch': 1.42} + 28%|██▊ | 213/750 [22:49<55:14, 6.17s/it] 29%|██▊ | 214/750 [22:55<54:32, 6.10s/it] {'loss': 0.2249, 'grad_norm': 0.10268723219633102, 'learning_rate': 0.0001639181468858914, 'epoch': 1.43} + 29%|██▊ | 214/750 [22:55<54:32, 6.10s/it] 29%|██▊ | 215/750 [23:01<54:22, 6.10s/it] {'loss': 0.1849, 'grad_norm': 0.09193839132785797, 'learning_rate': 0.00016353649759766664, 'epoch': 1.43} + 29%|██▊ | 215/750 [23:01<54:22, 6.10s/it] 29%|██▉ | 216/750 [23:07<54:06, 6.08s/it] {'loss': 0.1821, 'grad_norm': 0.09557466208934784, 'learning_rate': 0.00016315750172876014, 'epoch': 1.44} + 29%|██▉ | 216/750 [23:07<54:06, 6.08s/it] 29%|██▉ | 217/750 [23:13<53:50, 6.06s/it] {'loss': 0.1937, 'grad_norm': 0.09658323973417282, 'learning_rate': 0.00016278112867447063, 'epoch': 1.45} + 29%|██▉ | 217/750 [23:13<53:50, 6.06s/it] 29%|██▉ | 218/750 [23:19<53:15, 6.01s/it] {'loss': 0.1884, 'grad_norm': 0.09883001446723938, 'learning_rate': 0.00016240734832202275, 'epoch': 1.45} + 29%|██▉ | 218/750 [23:19<53:15, 6.01s/it] 29%|██▉ | 219/750 [23:24<51:57, 5.87s/it] {'loss': 0.1738, 'grad_norm': 0.09773606806993484, 'learning_rate': 0.00016203613104044751, 'epoch': 1.46} + 29%|██▉ | 219/750 [23:24<51:57, 5.87s/it] 29%|██▉ | 220/750 [23:30<52:14, 5.91s/it] {'loss': 0.1874, 'grad_norm': 0.09291878342628479, 'learning_rate': 0.00016166744767071581, 'epoch': 1.47} + 29%|██▉ | 220/750 [23:30<52:14, 5.91s/it] 29%|██▉ | 221/750 [23:36<51:34, 5.85s/it] {'loss': 0.1992, 'grad_norm': 0.11277790367603302, 'learning_rate': 0.00016130126951611793, 'epoch': 1.47} + 29%|██▉ | 221/750 [23:36<51:34, 5.85s/it] 30%|██▉ | 222/750 [23:42<52:28, 5.96s/it] {'loss': 0.1929, 'grad_norm': 0.1059861108660698, 'learning_rate': 0.0001609375683328815, 'epoch': 1.48} + 30%|██▉ | 222/750 [23:42<52:28, 5.96s/it] 30%|██▉ | 223/750 [23:48<52:27, 5.97s/it] {'loss': 0.2077, 'grad_norm': 0.0982198640704155, 'learning_rate': 0.00016057631632102133, 'epoch': 1.49} + 30%|██▉ | 223/750 [23:48<52:27, 5.97s/it] 30%|██▉ | 224/750 [23:54<52:08, 5.95s/it] {'loss': 0.1968, 'grad_norm': 0.10334406793117523, 'learning_rate': 0.00016021748611541394, 'epoch': 1.49} + 30%|██▉ | 224/750 [23:54<52:08, 5.95s/it] 30%|███ | 225/750 [24:00<52:08, 5.96s/it] {'loss': 0.1918, 'grad_norm': 0.09775006771087646, 'learning_rate': 0.00015986105077709064, 'epoch': 1.5} + 30%|███ | 225/750 [24:00<52:08, 5.96s/it] 30%|███ | 226/750 [24:06<52:11, 5.98s/it] {'loss': 0.2041, 'grad_norm': 0.09927330166101456, 'learning_rate': 0.00015950698378474278, 'epoch': 1.51} + 30%|███ | 226/750 [24:06<52:11, 5.98s/it] 30%|███ | 227/750 [24:12<52:08, 5.98s/it] {'loss': 0.1932, 'grad_norm': 0.09958826750516891, 'learning_rate': 0.00015915525902643283, 'epoch': 1.51} + 30%|███ | 227/750 [24:12<52:08, 5.98s/it] 30%|███ | 228/750 [24:18<51:38, 5.94s/it] {'loss': 0.1951, 'grad_norm': 0.10214662551879883, 'learning_rate': 0.0001588058507915059, 'epoch': 1.52} + 30%|███ | 228/750 [24:18<51:38, 5.94s/it] 31%|███ | 229/750 [24:24<51:23, 5.92s/it] {'loss': 0.192, 'grad_norm': 0.1043030396103859, 'learning_rate': 0.00015845873376269562, 'epoch': 1.53} + 31%|███ | 229/750 [24:24<51:23, 5.92s/it] 31%|███ | 230/750 [24:30<51:12, 5.91s/it] {'loss': 0.2051, 'grad_norm': 0.10158228129148483, 'learning_rate': 0.00015811388300841897, 'epoch': 1.53} + 31%|███ | 230/750 [24:30<51:12, 5.91s/it] 31%|███ | 231/750 [24:35<50:52, 5.88s/it] {'loss': 0.2134, 'grad_norm': 0.10783704370260239, 'learning_rate': 0.00015777127397525472, 'epoch': 1.54} + 31%|███ | 231/750 [24:35<50:52, 5.88s/it] 31%|███ | 232/750 [24:41<51:07, 5.92s/it] {'loss': 0.2005, 'grad_norm': 0.10326651483774185, 'learning_rate': 0.00015743088248060063, 'epoch': 1.55} + 31%|███ | 232/750 [24:41<51:07, 5.92s/it] 31%|███ | 233/750 [24:47<51:02, 5.92s/it] {'loss': 0.2053, 'grad_norm': 0.1080598458647728, 'learning_rate': 0.0001570926847055038, 'epoch': 1.55} + 31%|███ | 233/750 [24:47<51:02, 5.92s/it] 31%|███ | 234/750 [24:53<51:06, 5.94s/it] {'loss': 0.1944, 'grad_norm': 0.09971864521503448, 'learning_rate': 0.00015675665718766006, 'epoch': 1.56} + 31%|███ | 234/750 [24:53<51:06, 5.94s/it] 31%|███▏ | 235/750 [24:59<50:55, 5.93s/it] {'loss': 0.2076, 'grad_norm': 0.10496847331523895, 'learning_rate': 0.00015642277681457702, 'epoch': 1.57} + 31%|███▏ | 235/750 [24:59<50:55, 5.93s/it] 31%|███▏ | 236/750 [25:05<50:45, 5.93s/it] {'loss': 0.2104, 'grad_norm': 0.09720281511545181, 'learning_rate': 0.00015609102081689716, 'epoch': 1.57} + 31%|███▏ | 236/750 [25:05<50:45, 5.93s/it] 32%|███▏ | 237/750 [25:11<50:31, 5.91s/it] {'loss': 0.1897, 'grad_norm': 0.1059885248541832, 'learning_rate': 0.00015576136676187527, 'epoch': 1.58} + 32%|███▏ | 237/750 [25:11<50:31, 5.91s/it] 32%|███▏ | 238/750 [25:17<50:24, 5.91s/it] {'loss': 0.2077, 'grad_norm': 0.10878182202577591, 'learning_rate': 0.0001554337925470077, 'epoch': 1.59} + 32%|███▏ | 238/750 [25:17<50:24, 5.91s/it] 32%|███▏ | 239/750 [25:23<50:27, 5.92s/it] {'loss': 0.1978, 'grad_norm': 0.10656464844942093, 'learning_rate': 0.00015510827639380736, 'epoch': 1.59} + 32%|███▏ | 239/750 [25:23<50:27, 5.92s/it] 32%|███▏ | 240/750 [25:29<50:32, 5.95s/it] {'loss': 0.2059, 'grad_norm': 0.09944751858711243, 'learning_rate': 0.00015478479684172258, 'epoch': 1.6} + 32%|███▏ | 240/750 [25:29<50:32, 5.95s/it] 32%|███▏ | 241/750 [25:35<50:52, 6.00s/it] {'loss': 0.1813, 'grad_norm': 0.10171686112880707, 'learning_rate': 0.00015446333274219396, 'epoch': 1.61} + 32%|███▏ | 241/750 [25:35<50:52, 6.00s/it] 32%|███▏ | 242/750 [25:41<50:50, 6.00s/it] {'loss': 0.2015, 'grad_norm': 0.10150379687547684, 'learning_rate': 0.000154143863252847, 'epoch': 1.61} + 32%|███▏ | 242/750 [25:41<50:50, 6.00s/it] 32%|███▏ | 243/750 [25:47<50:20, 5.96s/it] {'loss': 0.19, 'grad_norm': 0.09964952617883682, 'learning_rate': 0.0001538263678318162, 'epoch': 1.62} + 32%|███▏ | 243/750 [25:47<50:20, 5.96s/it] 33%|███▎ | 244/750 [25:53<50:14, 5.96s/it] {'loss': 0.2034, 'grad_norm': 0.10983213782310486, 'learning_rate': 0.00015351082623219707, 'epoch': 1.63} + 33%|███▎ | 244/750 [25:53<50:14, 5.96s/it] 33%|███▎ | 245/750 [25:59<49:56, 5.93s/it] {'loss': 0.211, 'grad_norm': 0.10578607767820358, 'learning_rate': 0.00015319721849662297, 'epoch': 1.63} + 33%|███▎ | 245/750 [25:59<49:56, 5.93s/it] 33%|███▎ | 246/750 [26:05<49:55, 5.94s/it] {'loss': 0.2101, 'grad_norm': 0.1035049706697464, 'learning_rate': 0.0001528855249519633, 'epoch': 1.64} + 33%|███▎ | 246/750 [26:05<49:55, 5.94s/it] 33%|███▎ | 247/750 [26:11<49:56, 5.96s/it] {'loss': 0.2217, 'grad_norm': 0.09901197999715805, 'learning_rate': 0.00015257572620413985, 'epoch': 1.65} + 33%|███▎ | 247/750 [26:11<49:56, 5.96s/it] 33%|███▎ | 248/750 [26:17<49:48, 5.95s/it] {'loss': 0.1959, 'grad_norm': 0.09957315027713776, 'learning_rate': 0.00015226780313305814, 'epoch': 1.65} + 33%|███▎ | 248/750 [26:17<49:48, 5.95s/it] 33%|███▎ | 249/750 [26:23<49:47, 5.96s/it] {'loss': 0.1845, 'grad_norm': 0.10152138769626617, 'learning_rate': 0.00015196173688765093, 'epoch': 1.66} + 33%|███▎ | 249/750 [26:23<49:47, 5.96s/it] 33%|███▎ | 250/750 [26:29<49:34, 5.95s/it] {'loss': 0.1907, 'grad_norm': 0.10356403142213821, 'learning_rate': 0.00015165750888103104, 'epoch': 1.67} + 33%|███▎ | 250/750 [26:29<49:34, 5.95s/it] 33%|███▎ | 251/750 [26:34<49:13, 5.92s/it] {'loss': 0.1966, 'grad_norm': 0.10825596004724503, 'learning_rate': 0.00015135510078575002, 'epoch': 1.67} + 33%|███▎ | 251/750 [26:34<49:13, 5.92s/it] 34%|███▎ | 252/750 [26:40<49:12, 5.93s/it] {'loss': 0.1849, 'grad_norm': 0.101904958486557, 'learning_rate': 0.00015105449452916094, 'epoch': 1.68} + 34%|███▎ | 252/750 [26:40<49:12, 5.93s/it] 34%|███▎ | 253/750 [26:46<49:04, 5.93s/it] {'loss': 0.1913, 'grad_norm': 0.10389713197946548, 'learning_rate': 0.00015075567228888182, 'epoch': 1.69} + 34%|███▎ | 253/750 [26:46<49:04, 5.93s/it] 34%|███▍ | 254/750 [26:52<48:57, 5.92s/it] {'loss': 0.2084, 'grad_norm': 0.10135425627231598, 'learning_rate': 0.0001504586164883572, 'epoch': 1.69} + 34%|███▍ | 254/750 [26:52<48:57, 5.92s/it] 34%|███▍ | 255/750 [26:58<48:52, 5.92s/it] {'loss': 0.1956, 'grad_norm': 0.09966433793306351, 'learning_rate': 0.00015016330979251603, 'epoch': 1.7} + 34%|███▍ | 255/750 [26:58<48:52, 5.92s/it] 34%|███▍ | 256/750 [27:04<48:42, 5.92s/it] {'loss': 0.2018, 'grad_norm': 0.10264011472463608, 'learning_rate': 0.00014986973510352247, 'epoch': 1.71} + 34%|███▍ | 256/750 [27:04<48:42, 5.92s/it] 34%|███▍ | 257/750 [27:10<48:39, 5.92s/it] {'loss': 0.2023, 'grad_norm': 0.10318576544523239, 'learning_rate': 0.0001495778755566181, 'epoch': 1.71} + 34%|███▍ | 257/750 [27:10<48:39, 5.92s/it] 34%|███▍ | 258/750 [27:16<48:25, 5.91s/it] {'loss': 0.1893, 'grad_norm': 0.10010098665952682, 'learning_rate': 0.0001492877145160524, 'epoch': 1.72} + 34%|███▍ | 258/750 [27:16<48:25, 5.91s/it] 35%|███▍ | 259/750 [27:22<48:10, 5.89s/it] {'loss': 0.1984, 'grad_norm': 0.09545325487852097, 'learning_rate': 0.00014899923557110017, 'epoch': 1.73} + 35%|███▍ | 259/750 [27:22<48:10, 5.89s/it] 35%|███▍ | 260/750 [27:28<48:34, 5.95s/it] {'loss': 0.2001, 'grad_norm': 0.09963326156139374, 'learning_rate': 0.0001487124225321631, 'epoch': 1.73} + 35%|███▍ | 260/750 [27:28<48:34, 5.95s/it] 35%|███▍ | 261/750 [27:34<48:39, 5.97s/it] {'loss': 0.1866, 'grad_norm': 0.10346517711877823, 'learning_rate': 0.00014842725942695355, 'epoch': 1.74} + 35%|███▍ | 261/750 [27:34<48:39, 5.97s/it] 35%|███▍ | 262/750 [27:40<48:22, 5.95s/it] {'loss': 0.1709, 'grad_norm': 0.1005755364894867, 'learning_rate': 0.0001481437304967584, 'epoch': 1.75} + 35%|███▍ | 262/750 [27:40<48:22, 5.95s/it] 35%|███▌ | 263/750 [27:46<48:17, 5.95s/it] {'loss': 0.2003, 'grad_norm': 0.10314761102199554, 'learning_rate': 0.00014786182019278145, 'epoch': 1.75} + 35%|███▌ | 263/750 [27:46<48:17, 5.95s/it] 35%|███▌ | 264/750 [27:52<48:06, 5.94s/it] {'loss': 0.1899, 'grad_norm': 0.10142087191343307, 'learning_rate': 0.0001475815131725618, 'epoch': 1.76} + 35%|███▌ | 264/750 [27:52<48:06, 5.94s/it] 35%|███▌ | 265/750 [27:58<48:04, 5.95s/it] {'loss': 0.1996, 'grad_norm': 0.10087636858224869, 'learning_rate': 0.00014730279429646666, 'epoch': 1.77} + 35%|███▌ | 265/750 [27:58<48:04, 5.95s/it] 35%|███▌ | 266/750 [28:03<47:49, 5.93s/it] {'loss': 0.197, 'grad_norm': 0.10521330684423447, 'learning_rate': 0.0001470256486242569, 'epoch': 1.77} + 35%|███▌ | 266/750 [28:03<47:49, 5.93s/it] 36%|███▌ | 267/750 [28:09<47:36, 5.91s/it] {'loss': 0.192, 'grad_norm': 0.10543681681156158, 'learning_rate': 0.00014675006141172335, 'epoch': 1.78} + 36%|███▌ | 267/750 [28:09<47:36, 5.91s/it] 36%|███▌ | 268/750 [28:15<47:32, 5.92s/it] {'loss': 0.2002, 'grad_norm': 0.10189599543809891, 'learning_rate': 0.00014647601810739195, 'epoch': 1.79} + 36%|███▌ | 268/750 [28:15<47:32, 5.92s/it] 36%|███▌ | 269/750 [28:21<47:25, 5.92s/it] {'loss': 0.225, 'grad_norm': 0.11100862920284271, 'learning_rate': 0.0001462035043492969, 'epoch': 1.79} + 36%|███▌ | 269/750 [28:21<47:25, 5.92s/it] 36%|███▌ | 270/750 [28:27<47:29, 5.94s/it] {'loss': 0.1919, 'grad_norm': 0.09958995133638382, 'learning_rate': 0.00014593250596181887, 'epoch': 1.8} + 36%|███▌ | 270/750 [28:27<47:29, 5.94s/it] 36%|███▌ | 271/750 [28:33<47:16, 5.92s/it] {'loss': 0.1938, 'grad_norm': 0.10462497919797897, 'learning_rate': 0.00014566300895258813, 'epoch': 1.81} + 36%|███▌ | 271/750 [28:33<47:16, 5.92s/it] 36%|███▋ | 272/750 [28:39<47:06, 5.91s/it] {'loss': 0.1929, 'grad_norm': 0.11180876940488815, 'learning_rate': 0.00014539499950944992, 'epoch': 1.81} + 36%|███▋ | 272/750 [28:39<47:06, 5.91s/it] 36%|███▋ | 273/750 [28:45<47:10, 5.93s/it] {'loss': 0.1767, 'grad_norm': 0.0973215252161026, 'learning_rate': 0.00014512846399749107, 'epoch': 1.82} + 36%|███▋ | 273/750 [28:45<47:10, 5.93s/it] 37%|███▋ | 274/750 [28:51<47:18, 5.96s/it] {'loss': 0.1796, 'grad_norm': 0.09500745683908463, 'learning_rate': 0.00014486338895612657, 'epoch': 1.83} + 37%|███▋ | 274/750 [28:51<47:18, 5.96s/it] 37%|███▋ | 275/750 [28:57<46:55, 5.93s/it] {'loss': 0.1729, 'grad_norm': 0.10198456048965454, 'learning_rate': 0.00014459976109624422, 'epoch': 1.83} + 37%|███▋ | 275/750 [28:57<46:55, 5.93s/it] 37%|███▋ | 276/750 [29:03<46:44, 5.92s/it] {'loss': 0.1964, 'grad_norm': 0.1011200100183487, 'learning_rate': 0.00014433756729740648, 'epoch': 1.84} + 37%|███▋ | 276/750 [29:03<46:44, 5.92s/it] 37%|███▋ | 277/750 [29:08<46:24, 5.89s/it] {'loss': 0.2199, 'grad_norm': 0.11574704200029373, 'learning_rate': 0.00014407679460510741, 'epoch': 1.85} + 37%|███▋ | 277/750 [29:08<46:24, 5.89s/it] 37%|███▋ | 278/750 [29:14<46:22, 5.89s/it] {'loss': 0.1884, 'grad_norm': 0.10165125131607056, 'learning_rate': 0.00014381743022808492, 'epoch': 1.85} + 37%|███▋ | 278/750 [29:14<46:22, 5.89s/it] 37%|███▋ | 279/750 [29:20<46:31, 5.93s/it] {'loss': 0.1898, 'grad_norm': 0.09997145086526871, 'learning_rate': 0.00014355946153568483, 'epoch': 1.86} + 37%|███▋ | 279/750 [29:20<46:31, 5.93s/it] 37%|███▋ | 280/750 [29:26<46:39, 5.96s/it] {'loss': 0.2153, 'grad_norm': 0.11056888848543167, 'learning_rate': 0.0001433028760552777, 'epoch': 1.87} + 37%|███▋ | 280/750 [29:26<46:39, 5.96s/it] 37%|███▋ | 281/750 [29:32<46:19, 5.93s/it] {'loss': 0.2125, 'grad_norm': 0.10536083579063416, 'learning_rate': 0.00014304766146972588, 'epoch': 1.87} + 37%|███▋ | 281/750 [29:32<46:19, 5.93s/it] 38%|███▊ | 282/750 [29:38<46:15, 5.93s/it] {'loss': 0.1987, 'grad_norm': 0.10255546122789383, 'learning_rate': 0.0001427938056148999, 'epoch': 1.88} + 38%|███▊ | 282/750 [29:38<46:15, 5.93s/it] 38%|███▊ | 283/750 [29:44<46:06, 5.92s/it] {'loss': 0.1904, 'grad_norm': 0.10317458212375641, 'learning_rate': 0.00014254129647724352, 'epoch': 1.89} + 38%|███▊ | 283/750 [29:44<46:06, 5.92s/it] 38%|███▊ | 284/750 [29:50<46:01, 5.93s/it] {'loss': 0.2045, 'grad_norm': 0.1047426164150238, 'learning_rate': 0.00014229012219138556, 'epoch': 1.89} + 38%|███▊ | 284/750 [29:50<46:01, 5.93s/it] 38%|███▊ | 285/750 [29:56<45:52, 5.92s/it] {'loss': 0.2055, 'grad_norm': 0.09543642401695251, 'learning_rate': 0.00014204027103779837, 'epoch': 1.9} + 38%|███▊ | 285/750 [29:56<45:52, 5.92s/it] 38%|███▊ | 286/750 [30:02<45:53, 5.93s/it] {'loss': 0.2, 'grad_norm': 0.10057047009468079, 'learning_rate': 0.0001417917314405008, 'epoch': 1.91} + 38%|███▊ | 286/750 [30:02<45:53, 5.93s/it] 38%|███▊ | 287/750 [30:08<45:44, 5.93s/it] {'loss': 0.1937, 'grad_norm': 0.10736412554979324, 'learning_rate': 0.0001415444919648056, 'epoch': 1.91} + 38%|███▊ | 287/750 [30:08<45:44, 5.93s/it] 38%|███▊ | 288/750 [30:14<45:42, 5.94s/it] {'loss': 0.2036, 'grad_norm': 0.10363559424877167, 'learning_rate': 0.00014129854131510975, 'epoch': 1.92} + 38%|███▊ | 288/750 [30:14<45:42, 5.94s/it] 39%|███▊ | 289/750 [30:20<45:36, 5.94s/it] {'loss': 0.1796, 'grad_norm': 0.10259436815977097, 'learning_rate': 0.00014105386833272704, 'epoch': 1.93} + 39%|███▊ | 289/750 [30:20<45:36, 5.94s/it] 39%|███▊ | 290/750 [30:26<45:31, 5.94s/it] {'loss': 0.2107, 'grad_norm': 0.1215776577591896, 'learning_rate': 0.0001408104619937615, 'epoch': 1.93} + 39%|███▊ | 290/750 [30:26<45:31, 5.94s/it] 39%|███▉ | 291/750 [30:31<45:10, 5.91s/it] {'loss': 0.1842, 'grad_norm': 0.11006634682416916, 'learning_rate': 0.00014056831140702128, 'epoch': 1.94} + 39%|███▉ | 291/750 [30:31<45:10, 5.91s/it] 39%|███▉ | 292/750 [30:37<45:05, 5.91s/it] {'loss': 0.2124, 'grad_norm': 0.10849900543689728, 'learning_rate': 0.0001403274058119718, 'epoch': 1.95} + 39%|███▉ | 292/750 [30:37<45:05, 5.91s/it] 39%|███▉ | 293/750 [30:43<45:00, 5.91s/it] {'loss': 0.2102, 'grad_norm': 0.11632896959781647, 'learning_rate': 0.00014008773457672727, 'epoch': 1.95} + 39%|███▉ | 293/750 [30:43<45:00, 5.91s/it] 39%|███▉ | 294/750 [30:49<44:48, 5.90s/it] {'loss': 0.1734, 'grad_norm': 0.10694096237421036, 'learning_rate': 0.00013984928719608012, 'epoch': 1.96} + 39%|███▉ | 294/750 [30:49<44:48, 5.90s/it] 39%|███▉ | 295/750 [30:55<44:48, 5.91s/it] {'loss': 0.2108, 'grad_norm': 0.11361347138881683, 'learning_rate': 0.0001396120532895667, 'epoch': 1.97} + 39%|███▉ | 295/750 [30:55<44:48, 5.91s/it] 39%|███▉ | 296/750 [31:01<44:57, 5.94s/it] {'loss': 0.21, 'grad_norm': 0.1309555172920227, 'learning_rate': 0.00013937602259956938, 'epoch': 1.97} + 39%|███▉ | 296/750 [31:01<44:57, 5.94s/it] 40%|███▉ | 297/750 [31:07<45:19, 6.00s/it] {'loss': 0.1977, 'grad_norm': 0.1072709858417511, 'learning_rate': 0.00013914118498945362, 'epoch': 1.98} + 40%|███▉ | 297/750 [31:07<45:19, 6.00s/it] 40%|███▉ | 298/750 [31:13<45:07, 5.99s/it] {'loss': 0.1671, 'grad_norm': 0.10048194974660873, 'learning_rate': 0.00013890753044173962, 'epoch': 1.99} + 40%|███▉ | 298/750 [31:13<45:07, 5.99s/it] 40%|███▉ | 299/750 [31:19<45:07, 6.00s/it] {'loss': 0.1928, 'grad_norm': 0.10817109793424606, 'learning_rate': 0.0001386750490563073, 'epoch': 1.99} + 40%|███▉ | 299/750 [31:19<45:07, 6.00s/it] 40%|████ | 300/750 [31:27<49:42, 6.63s/it] {'loss': 0.1744, 'grad_norm': 0.10435310751199722, 'learning_rate': 0.00013844373104863458, 'epoch': 2.0} + 40%|████ | 300/750 [31:27<49:42, 6.63s/it] 40%|████ | 301/750 [31:42<1:07:11, 8.98s/it] {'loss': 0.1826, 'grad_norm': 0.10341206192970276, 'learning_rate': 0.00013821356674806781, 'epoch': 2.01} + 40%|████ | 301/750 [31:42<1:07:11, 8.98s/it] 40%|████ | 302/750 [31:48<1:00:09, 8.06s/it] {'loss': 0.1903, 'grad_norm': 0.10279000550508499, 'learning_rate': 0.00013798454659612338, 'epoch': 2.01} + 40%|████ | 302/750 [31:48<1:00:09, 8.06s/it] 40%|████ | 303/750 [31:53<54:51, 7.36s/it] {'loss': 0.1545, 'grad_norm': 0.10367275774478912, 'learning_rate': 0.00013775666114482078, 'epoch': 2.02} + 40%|████ | 303/750 [31:53<54:51, 7.36s/it] 41%|████ | 304/750 [31:59<51:38, 6.95s/it] {'loss': 0.1745, 'grad_norm': 0.10173984616994858, 'learning_rate': 0.0001375299010550452, 'epoch': 2.03} + 41%|████ | 304/750 [31:59<51:38, 6.95s/it] 41%|████ | 305/750 [32:05<49:18, 6.65s/it] {'loss': 0.173, 'grad_norm': 0.10839492827653885, 'learning_rate': 0.0001373042570949402, 'epoch': 2.03} + 41%|████ | 305/750 [32:05<49:18, 6.65s/it] 41%|████ | 306/750 [32:11<47:39, 6.44s/it] {'loss': 0.1693, 'grad_norm': 0.10868673026561737, 'learning_rate': 0.00013707972013832906, 'epoch': 2.04} + 41%|████ | 306/750 [32:11<47:39, 6.44s/it] 41%|████ | 307/750 [32:17<46:29, 6.30s/it] {'loss': 0.1621, 'grad_norm': 0.11517327278852463, 'learning_rate': 0.0001368562811631643, 'epoch': 2.05} + 41%|████ | 307/750 [32:17<46:29, 6.30s/it] 41%|████ | 308/750 [32:23<45:48, 6.22s/it] {'loss': 0.1666, 'grad_norm': 0.10989931225776672, 'learning_rate': 0.0001366339312500053, 'epoch': 2.05} + 41%|████ | 308/750 [32:23<45:48, 6.22s/it] 41%|████ | 309/750 [32:29<45:01, 6.13s/it] {'loss': 0.1804, 'grad_norm': 0.13048018515110016, 'learning_rate': 0.00013641266158052266, 'epoch': 2.06} + 41%|████ | 309/750 [32:29<45:01, 6.13s/it] 41%|████▏ | 310/750 [32:35<44:25, 6.06s/it] {'loss': 0.1862, 'grad_norm': 0.1161588802933693, 'learning_rate': 0.00013619246343602937, 'epoch': 2.07} + 41%|████▏ | 310/750 [32:35<44:25, 6.06s/it] 41%|████▏ | 311/750 [32:41<44:06, 6.03s/it] {'loss': 0.1887, 'grad_norm': 0.13126613199710846, 'learning_rate': 0.00013597332819603774, 'epoch': 2.07} + 41%|████▏ | 311/750 [32:41<44:06, 6.03s/it] 42%|████▏ | 312/750 [32:47<43:37, 5.98s/it] {'loss': 0.1845, 'grad_norm': 0.12416185438632965, 'learning_rate': 0.00013575524733684213, 'epoch': 2.08} + 42%|████▏ | 312/750 [32:47<43:37, 5.98s/it] 42%|████▏ | 313/750 [32:53<43:16, 5.94s/it] {'loss': 0.1608, 'grad_norm': 0.11586000770330429, 'learning_rate': 0.00013553821243012613, 'epoch': 2.09} + 42%|████▏ | 313/750 [32:53<43:16, 5.94s/it] 42%|████▏ | 314/750 [32:59<43:18, 5.96s/it] {'loss': 0.1742, 'grad_norm': 0.11065978556871414, 'learning_rate': 0.00013532221514159462, 'epoch': 2.09} + 42%|████▏ | 314/750 [32:59<43:18, 5.96s/it] 42%|████▏ | 315/750 [33:05<43:02, 5.94s/it] {'loss': 0.1823, 'grad_norm': 0.12366791069507599, 'learning_rate': 0.00013510724722962957, 'epoch': 2.1} + 42%|████▏ | 315/750 [33:05<43:02, 5.94s/it] 42%|████▏ | 316/750 [33:11<43:06, 5.96s/it] {'loss': 0.1739, 'grad_norm': 0.11502339690923691, 'learning_rate': 0.00013489330054396908, 'epoch': 2.11} + 42%|████▏ | 316/750 [33:11<43:06, 5.96s/it] 42%|████▏ | 317/750 [33:17<43:35, 6.04s/it] {'loss': 0.1851, 'grad_norm': 0.12158779054880142, 'learning_rate': 0.0001346803670244095, 'epoch': 2.11} + 42%|████▏ | 317/750 [33:17<43:35, 6.04s/it] 42%|████▏ | 318/750 [33:23<43:36, 6.06s/it] {'loss': 0.168, 'grad_norm': 0.1147841066122055, 'learning_rate': 0.00013446843869953018, 'epoch': 2.12} + 42%|████▏ | 318/750 [33:23<43:36, 6.06s/it] 43%|████▎ | 319/750 [33:29<43:08, 6.01s/it] {'loss': 0.158, 'grad_norm': 0.11651469022035599, 'learning_rate': 0.0001342575076854399, 'epoch': 2.13} + 43%|████▎ | 319/750 [33:29<43:08, 6.01s/it] 43%|████▎ | 320/750 [33:35<42:49, 5.98s/it] {'loss': 0.1726, 'grad_norm': 0.11769665032625198, 'learning_rate': 0.00013404756618454512, 'epoch': 2.13} + 43%|████▎ | 320/750 [33:35<42:49, 5.98s/it] 43%|████▎ | 321/750 [33:41<42:12, 5.90s/it] {'loss': 0.1777, 'grad_norm': 0.12741614878177643, 'learning_rate': 0.00013383860648433912, 'epoch': 2.14} + 43%|████▎ | 321/750 [33:41<42:12, 5.90s/it] 43%|████▎ | 322/750 [33:47<42:13, 5.92s/it] {'loss': 0.1847, 'grad_norm': 0.12269234657287598, 'learning_rate': 0.0001336306209562122, 'epoch': 2.15} + 43%|████▎ | 322/750 [33:47<42:13, 5.92s/it] 43%|████▎ | 323/750 [33:52<42:06, 5.92s/it] {'loss': 0.1649, 'grad_norm': 0.1309356391429901, 'learning_rate': 0.0001334236020542817, 'epoch': 2.15} + 43%|████▎ | 323/750 [33:52<42:06, 5.92s/it] 43%|████▎ | 324/750 [33:58<42:01, 5.92s/it] {'loss': 0.1771, 'grad_norm': 0.127553790807724, 'learning_rate': 0.00013321754231424221, 'epoch': 2.16} + 43%|████▎ | 324/750 [33:58<42:01, 5.92s/it] 43%|████▎ | 325/750 [34:04<41:54, 5.92s/it] {'loss': 0.1811, 'grad_norm': 0.12544496357440948, 'learning_rate': 0.00013301243435223526, 'epoch': 2.17} + 43%|████▎ | 325/750 [34:04<41:54, 5.92s/it] 43%|████▎ | 326/750 [34:10<41:56, 5.93s/it] {'loss': 0.1734, 'grad_norm': 0.11943470686674118, 'learning_rate': 0.00013280827086373782, 'epoch': 2.17} + 43%|████▎ | 326/750 [34:10<41:56, 5.93s/it] 44%|████▎ | 327/750 [34:16<41:41, 5.91s/it] {'loss': 0.1784, 'grad_norm': 0.12659519910812378, 'learning_rate': 0.00013260504462246983, 'epoch': 2.18} + 44%|████▎ | 327/750 [34:16<41:41, 5.91s/it] 44%|████▎ | 328/750 [34:22<41:48, 5.94s/it] {'loss': 0.1457, 'grad_norm': 0.11461290717124939, 'learning_rate': 0.00013240274847931992, 'epoch': 2.19} + 44%|████▎ | 328/750 [34:22<41:48, 5.94s/it] 44%|████▍ | 329/750 [34:28<41:43, 5.95s/it] {'loss': 0.1749, 'grad_norm': 0.12830543518066406, 'learning_rate': 0.00013220137536128893, 'epoch': 2.19} + 44%|████▍ | 329/750 [34:28<41:43, 5.95s/it] 44%|████▍ | 330/750 [34:34<41:39, 5.95s/it] {'loss': 0.1542, 'grad_norm': 0.12447745352983475, 'learning_rate': 0.00013200091827045155, 'epoch': 2.2} + 44%|████▍ | 330/750 [34:34<41:39, 5.95s/it] 44%|████▍ | 331/750 [34:40<41:32, 5.95s/it] {'loss': 0.1701, 'grad_norm': 0.12327922135591507, 'learning_rate': 0.00013180137028293452, 'epoch': 2.21} + 44%|████▍ | 331/750 [34:40<41:32, 5.95s/it] 44%|████▍ | 332/750 [34:46<41:12, 5.91s/it] {'loss': 0.1665, 'grad_norm': 0.13163968920707703, 'learning_rate': 0.0001316027245479125, 'epoch': 2.21} + 44%|████▍ | 332/750 [34:46<41:12, 5.91s/it] 44%|████▍ | 333/750 [34:52<41:06, 5.92s/it] {'loss': 0.1551, 'grad_norm': 0.11950532346963882, 'learning_rate': 0.00013140497428662003, 'epoch': 2.22} + 44%|████▍ | 333/750 [34:52<41:06, 5.92s/it] 45%|████▍ | 334/750 [34:58<40:45, 5.88s/it] {'loss': 0.1987, 'grad_norm': 0.13608191907405853, 'learning_rate': 0.00013120811279137981, 'epoch': 2.23} + 45%|████▍ | 334/750 [34:58<40:45, 5.88s/it] 45%|████▍ | 335/750 [35:04<41:09, 5.95s/it] {'loss': 0.1631, 'grad_norm': 0.12113139778375626, 'learning_rate': 0.0001310121334246474, 'epoch': 2.23} + 45%|████▍ | 335/750 [35:04<41:09, 5.95s/it] 45%|████▍ | 336/750 [35:10<41:03, 5.95s/it] {'loss': 0.1422, 'grad_norm': 0.11814668029546738, 'learning_rate': 0.0001308170296180709, 'epoch': 2.24} + 45%|████▍ | 336/750 [35:10<41:03, 5.95s/it] 45%|████▍ | 337/750 [35:16<41:13, 5.99s/it] {'loss': 0.1691, 'grad_norm': 0.1233840212225914, 'learning_rate': 0.00013062279487156654, 'epoch': 2.25} + 45%|████▍ | 337/750 [35:16<41:13, 5.99s/it] 45%|████▌ | 338/750 [35:22<40:58, 5.97s/it] {'loss': 0.179, 'grad_norm': 0.13394248485565186, 'learning_rate': 0.00013042942275240903, 'epoch': 2.25} + 45%|████▌ | 338/750 [35:22<40:58, 5.97s/it] 45%|████▌ | 339/750 [35:28<40:47, 5.96s/it] {'loss': 0.1764, 'grad_norm': 0.12861736118793488, 'learning_rate': 0.0001302369068943367, 'epoch': 2.26} + 45%|████▌ | 339/750 [35:28<40:47, 5.96s/it] 45%|████▌ | 340/750 [35:33<40:41, 5.96s/it] {'loss': 0.1584, 'grad_norm': 0.12160958349704742, 'learning_rate': 0.00013004524099667142, 'epoch': 2.27} + 45%|████▌ | 340/750 [35:33<40:41, 5.96s/it] 45%|████▌ | 341/750 [35:39<40:40, 5.97s/it] {'loss': 0.1652, 'grad_norm': 0.12228940427303314, 'learning_rate': 0.00012985441882345223, 'epoch': 2.27} + 45%|████▌ | 341/750 [35:39<40:40, 5.97s/it] 46%|████▌ | 342/750 [35:45<40:35, 5.97s/it] {'loss': 0.1755, 'grad_norm': 0.13051892817020416, 'learning_rate': 0.00012966443420258316, 'epoch': 2.28} + 46%|████▌ | 342/750 [35:45<40:35, 5.97s/it] 46%|████▌ | 343/750 [35:51<40:22, 5.95s/it] {'loss': 0.1669, 'grad_norm': 0.12235318124294281, 'learning_rate': 0.00012947528102499512, 'epoch': 2.29} + 46%|████▌ | 343/750 [35:51<40:22, 5.95s/it] 46%|████▌ | 344/750 [35:57<40:17, 5.96s/it] {'loss': 0.1706, 'grad_norm': 0.12630563974380493, 'learning_rate': 0.00012928695324382025, 'epoch': 2.29} + 46%|████▌ | 344/750 [35:57<40:17, 5.96s/it] 46%|████▌ | 345/750 [36:03<39:59, 5.93s/it] {'loss': 0.1758, 'grad_norm': 0.1292877495288849, 'learning_rate': 0.00012909944487358055, 'epoch': 2.3} + 46%|████▌ | 345/750 [36:03<39:59, 5.93s/it] 46%|████▌ | 346/750 [36:09<39:45, 5.91s/it] {'loss': 0.1732, 'grad_norm': 0.12853533029556274, 'learning_rate': 0.0001289127499893885, 'epoch': 2.31} + 46%|████▌ | 346/750 [36:09<39:45, 5.91s/it] 46%|████▋ | 347/750 [36:15<39:56, 5.95s/it] {'loss': 0.1696, 'grad_norm': 0.12569411098957062, 'learning_rate': 0.00012872686272616092, 'epoch': 2.31} + 46%|████▋ | 347/750 [36:15<39:56, 5.95s/it] 46%|████▋ | 348/750 [36:21<39:45, 5.93s/it] {'loss': 0.1798, 'grad_norm': 0.12562617659568787, 'learning_rate': 0.0001285417772778451, 'epoch': 2.32} + 46%|████▋ | 348/750 [36:21<39:45, 5.93s/it] 47%|████▋ | 349/750 [36:27<39:35, 5.92s/it] {'loss': 0.1729, 'grad_norm': 0.12314952164888382, 'learning_rate': 0.0001283574878966568, 'epoch': 2.33} + 47%|████▋ | 349/750 [36:27<39:35, 5.92s/it] 47%|████▋ | 350/750 [36:33<39:39, 5.95s/it] {'loss': 0.1605, 'grad_norm': 0.11700372397899628, 'learning_rate': 0.00012817398889233115, 'epoch': 2.33} + 47%|████▋ | 350/750 [36:33<39:39, 5.95s/it] 47%|████▋ | 351/750 [36:39<39:40, 5.97s/it] {'loss': 0.1685, 'grad_norm': 0.12326090782880783, 'learning_rate': 0.00012799127463138408, 'epoch': 2.34} + 47%|████▋ | 351/750 [36:39<39:40, 5.97s/it] 47%|████▋ | 352/750 [36:45<39:20, 5.93s/it] {'loss': 0.1725, 'grad_norm': 0.13502374291419983, 'learning_rate': 0.00012780933953638628, 'epoch': 2.35} + 47%|████▋ | 352/750 [36:45<39:20, 5.93s/it] 47%|████▋ | 353/750 [36:51<39:22, 5.95s/it] {'loss': 0.1696, 'grad_norm': 0.12666338682174683, 'learning_rate': 0.00012762817808524798, 'epoch': 2.35} + 47%|████▋ | 353/750 [36:51<39:22, 5.95s/it] 47%|████▋ | 354/750 [36:57<39:38, 6.01s/it] {'loss': 0.1824, 'grad_norm': 0.13442446291446686, 'learning_rate': 0.00012744778481051504, 'epoch': 2.36} + 47%|████▋ | 354/750 [36:57<39:38, 6.01s/it] 47%|████▋ | 355/750 [37:03<39:27, 5.99s/it] {'loss': 0.1508, 'grad_norm': 0.1286042481660843, 'learning_rate': 0.0001272681542986758, 'epoch': 2.37} + 47%|████▋ | 355/750 [37:03<39:27, 5.99s/it] 47%|████▋ | 356/750 [37:09<39:17, 5.98s/it] {'loss': 0.1776, 'grad_norm': 0.13393321633338928, 'learning_rate': 0.0001270892811894789, 'epoch': 2.37} + 47%|████▋ | 356/750 [37:09<39:17, 5.98s/it] 48%|████▊ | 357/750 [37:15<39:08, 5.98s/it] {'loss': 0.1716, 'grad_norm': 0.13838514685630798, 'learning_rate': 0.00012691116017526124, 'epoch': 2.38} + 48%|████▊ | 357/750 [37:15<39:08, 5.98s/it] 48%|████▊ | 358/750 [37:21<39:03, 5.98s/it] {'loss': 0.1916, 'grad_norm': 0.1329745650291443, 'learning_rate': 0.00012673378600028677, 'epoch': 2.39} + 48%|████▊ | 358/750 [37:21<39:03, 5.98s/it] 48%|████▊ | 359/750 [37:27<38:42, 5.94s/it] {'loss': 0.1953, 'grad_norm': 0.13556192815303802, 'learning_rate': 0.00012655715346009488, 'epoch': 2.39} + 48%|████▊ | 359/750 [37:27<38:42, 5.94s/it] 48%|████▊ | 360/750 [37:33<38:37, 5.94s/it] {'loss': 0.1852, 'grad_norm': 0.12661460041999817, 'learning_rate': 0.00012638125740085918, 'epoch': 2.4} + 48%|████▊ | 360/750 [37:33<38:37, 5.94s/it] 48%|████▊ | 361/750 [37:38<38:29, 5.94s/it] {'loss': 0.168, 'grad_norm': 0.13118870556354523, 'learning_rate': 0.0001262060927187558, 'epoch': 2.41} + 48%|████▊ | 361/750 [37:38<38:29, 5.94s/it] 48%|████▊ | 362/750 [37:44<38:25, 5.94s/it] {'loss': 0.1705, 'grad_norm': 0.1364632397890091, 'learning_rate': 0.00012603165435934137, 'epoch': 2.41} + 48%|████▊ | 362/750 [37:44<38:25, 5.94s/it] 48%|████▊ | 363/750 [37:50<38:21, 5.95s/it] {'loss': 0.1681, 'grad_norm': 0.12689675390720367, 'learning_rate': 0.00012585793731694053, 'epoch': 2.42} + 48%|████▊ | 363/750 [37:50<38:21, 5.95s/it] 49%|████▊ | 364/750 [37:56<38:10, 5.93s/it] {'loss': 0.1613, 'grad_norm': 0.1276135891675949, 'learning_rate': 0.00012568493663404258, 'epoch': 2.43} + 49%|████▊ | 364/750 [37:56<38:10, 5.93s/it] 49%|████▊ | 365/750 [38:02<38:07, 5.94s/it] {'loss': 0.1784, 'grad_norm': 0.13087023794651031, 'learning_rate': 0.00012551264740070717, 'epoch': 2.43} + 49%|████▊ | 365/750 [38:02<38:07, 5.94s/it] 49%|████▉ | 366/750 [38:08<38:02, 5.94s/it] {'loss': 0.1973, 'grad_norm': 0.1365162432193756, 'learning_rate': 0.0001253410647539792, 'epoch': 2.44} + 49%|████▉ | 366/750 [38:08<38:02, 5.94s/it] 49%|████▉ | 367/750 [38:14<37:51, 5.93s/it] {'loss': 0.1468, 'grad_norm': 0.12326237559318542, 'learning_rate': 0.00012517018387731205, 'epoch': 2.45} + 49%|████▉ | 367/750 [38:14<37:51, 5.93s/it] 49%|████▉ | 368/750 [38:20<37:46, 5.93s/it] {'loss': 0.1823, 'grad_norm': 0.14036639034748077, 'learning_rate': 0.000125, 'epoch': 2.45} + 49%|████▉ | 368/750 [38:20<37:46, 5.93s/it] 49%|████▉ | 369/750 [38:26<37:44, 5.94s/it] {'loss': 0.1627, 'grad_norm': 0.12352321296930313, 'learning_rate': 0.0001248305083966186, 'epoch': 2.46} + 49%|████▉ | 369/750 [38:26<37:44, 5.94s/it] 49%|████▉ | 370/750 [38:32<37:32, 5.93s/it] {'loss': 0.1674, 'grad_norm': 0.13585619628429413, 'learning_rate': 0.00012466170438647366, 'epoch': 2.47} + 49%|████▉ | 370/750 [38:32<37:32, 5.93s/it] 49%|████▉ | 371/750 [38:38<37:20, 5.91s/it] {'loss': 0.1677, 'grad_norm': 0.12937146425247192, 'learning_rate': 0.00012449358333305828, 'epoch': 2.47} + 49%|████▉ | 371/750 [38:38<37:20, 5.91s/it] 50%|████▉ | 372/750 [38:44<37:12, 5.91s/it] {'loss': 0.1858, 'grad_norm': 0.1331632137298584, 'learning_rate': 0.00012432614064351803, 'epoch': 2.48} + 50%|████▉ | 372/750 [38:44<37:12, 5.91s/it] 50%|████▉ | 373/750 [38:50<37:09, 5.91s/it] {'loss': 0.1632, 'grad_norm': 0.13273197412490845, 'learning_rate': 0.00012415937176812403, 'epoch': 2.49} + 50%|████▉ | 373/750 [38:50<37:09, 5.91s/it] 50%|████▉ | 374/750 [38:56<37:14, 5.94s/it] {'loss': 0.166, 'grad_norm': 0.12437647581100464, 'learning_rate': 0.0001239932721997534, 'epoch': 2.49} + 50%|████▉ | 374/750 [38:56<37:14, 5.94s/it] 50%|█████ | 375/750 [39:02<37:09, 5.94s/it] {'loss': 0.1509, 'grad_norm': 0.12186422199010849, 'learning_rate': 0.00012382783747337805, 'epoch': 2.5} + 50%|█████ | 375/750 [39:02<37:09, 5.94s/it] 50%|█████ | 376/750 [39:08<37:05, 5.95s/it] {'loss': 0.1969, 'grad_norm': 0.1394951343536377, 'learning_rate': 0.0001236630631655603, 'epoch': 2.51} + 50%|█████ | 376/750 [39:08<37:05, 5.95s/it] 50%|█████ | 377/750 [39:13<36:56, 5.94s/it] {'loss': 0.1672, 'grad_norm': 0.1304289847612381, 'learning_rate': 0.0001234989448939562, 'epoch': 2.51} + 50%|█████ | 377/750 [39:13<36:56, 5.94s/it] 50%|█████ | 378/750 [39:19<36:51, 5.94s/it] {'loss': 0.1866, 'grad_norm': 0.13733752071857452, 'learning_rate': 0.00012333547831682581, 'epoch': 2.52} + 50%|█████ | 378/750 [39:19<36:51, 5.94s/it] 51%|█████ | 379/750 [39:25<36:39, 5.93s/it] {'loss': 0.16, 'grad_norm': 0.1302674114704132, 'learning_rate': 0.00012317265913255117, 'epoch': 2.53} + 51%|█████ | 379/750 [39:25<36:39, 5.93s/it] 51%|█████ | 380/750 [39:31<36:29, 5.92s/it] {'loss': 0.1706, 'grad_norm': 0.1256447583436966, 'learning_rate': 0.00012301048307916047, 'epoch': 2.53} + 51%|█████ | 380/750 [39:31<36:29, 5.92s/it] 51%|█████ | 381/750 [39:37<36:29, 5.93s/it] {'loss': 0.1636, 'grad_norm': 0.12781690061092377, 'learning_rate': 0.00012284894593385964, 'epoch': 2.54} + 51%|█████ | 381/750 [39:37<36:29, 5.93s/it] 51%|█████ | 382/750 [39:43<36:13, 5.91s/it] {'loss': 0.1725, 'grad_norm': 0.13997279107570648, 'learning_rate': 0.00012268804351257058, 'epoch': 2.55} + 51%|█████ | 382/750 [39:43<36:13, 5.91s/it] 51%|█████ | 383/750 [39:49<36:13, 5.92s/it] {'loss': 0.1798, 'grad_norm': 0.13648025691509247, 'learning_rate': 0.00012252777166947586, 'epoch': 2.55} + 51%|█████ | 383/750 [39:49<36:13, 5.92s/it] 51%|█████ | 384/750 [39:55<36:10, 5.93s/it] {'loss': 0.1619, 'grad_norm': 0.12897974252700806, 'learning_rate': 0.0001223681262965701, 'epoch': 2.56} + 51%|█████ | 384/750 [39:55<36:10, 5.93s/it] 51%|█████▏ | 385/750 [40:01<36:10, 5.95s/it] {'loss': 0.1682, 'grad_norm': 0.126958429813385, 'learning_rate': 0.00012220910332321784, 'epoch': 2.57} + 51%|█████▏ | 385/750 [40:01<36:10, 5.95s/it] 51%|█████▏ | 386/750 [40:07<35:57, 5.93s/it] {'loss': 0.1846, 'grad_norm': 0.14244182407855988, 'learning_rate': 0.00012205069871571739, 'epoch': 2.57} + 51%|█████▏ | 386/750 [40:07<35:57, 5.93s/it] 52%|█████▏ | 387/750 [40:13<35:52, 5.93s/it] {'loss': 0.1741, 'grad_norm': 0.13084439933300018, 'learning_rate': 0.00012189290847687124, 'epoch': 2.58} + 52%|█████▏ | 387/750 [40:13<35:52, 5.93s/it] 52%|█████▏ | 388/750 [40:19<35:49, 5.94s/it] {'loss': 0.1603, 'grad_norm': 0.12572620809078217, 'learning_rate': 0.0001217357286455623, 'epoch': 2.59} + 52%|█████▏ | 388/750 [40:19<35:49, 5.94s/it] 52%|█████▏ | 389/750 [40:25<35:48, 5.95s/it] {'loss': 0.1558, 'grad_norm': 0.12559637427330017, 'learning_rate': 0.00012157915529633644, 'epoch': 2.59} + 52%|█████▏ | 389/750 [40:25<35:48, 5.95s/it] 52%|█████▏ | 390/750 [40:31<35:54, 5.98s/it] {'loss': 0.1539, 'grad_norm': 0.12262400984764099, 'learning_rate': 0.00012142318453899053, 'epoch': 2.6} + 52%|█████▏ | 390/750 [40:31<35:54, 5.98s/it] 52%|█████▏ | 391/750 [40:37<35:36, 5.95s/it] {'loss': 0.1708, 'grad_norm': 0.13116472959518433, 'learning_rate': 0.00012126781251816648, 'epoch': 2.61} + 52%|█████▏ | 391/750 [40:37<35:36, 5.95s/it] 52%|█████▏ | 392/750 [40:42<35:23, 5.93s/it] {'loss': 0.1699, 'grad_norm': 0.13885435461997986, 'learning_rate': 0.00012111303541295122, 'epoch': 2.61} + 52%|█████▏ | 392/750 [40:42<35:23, 5.93s/it] 52%|█████▏ | 393/750 [40:48<35:25, 5.95s/it] {'loss': 0.1796, 'grad_norm': 0.1409914493560791, 'learning_rate': 0.00012095884943648174, 'epoch': 2.62} + 52%|█████▏ | 393/750 [40:48<35:25, 5.95s/it] 53%|█████▎ | 394/750 [40:54<35:10, 5.93s/it] {'loss': 0.1657, 'grad_norm': 0.12736555933952332, 'learning_rate': 0.0001208052508355561, 'epoch': 2.63} + 53%|█████▎ | 394/750 [40:54<35:10, 5.93s/it] 53%|█████▎ | 395/750 [41:00<35:11, 5.95s/it] {'loss': 0.1533, 'grad_norm': 0.11867203563451767, 'learning_rate': 0.0001206522358902497, 'epoch': 2.63} + 53%|█████▎ | 395/750 [41:00<35:11, 5.95s/it] 53%|█████▎ | 396/750 [41:06<34:55, 5.92s/it] {'loss': 0.145, 'grad_norm': 0.13425129652023315, 'learning_rate': 0.00012049980091353687, 'epoch': 2.64} + 53%|█████▎ | 396/750 [41:06<34:55, 5.92s/it] 53%|█████▎ | 397/750 [41:12<34:43, 5.90s/it] {'loss': 0.1727, 'grad_norm': 0.13228760659694672, 'learning_rate': 0.00012034794225091773, 'epoch': 2.65} + 53%|█████▎ | 397/750 [41:12<34:43, 5.90s/it] 53%|█████▎ | 398/750 [41:18<34:40, 5.91s/it] {'loss': 0.1629, 'grad_norm': 0.12948517501354218, 'learning_rate': 0.00012019665628005017, 'epoch': 2.65} + 53%|█████▎ | 398/750 [41:18<34:40, 5.91s/it] 53%|█████▎ | 399/750 [41:24<34:47, 5.95s/it] {'loss': 0.1785, 'grad_norm': 0.12350346148014069, 'learning_rate': 0.00012004593941038698, 'epoch': 2.66} + 53%|█████▎ | 399/750 [41:24<34:47, 5.95s/it] 53%|█████▎ | 400/750 [41:30<34:42, 5.95s/it] {'loss': 0.167, 'grad_norm': 0.1273069828748703, 'learning_rate': 0.00011989578808281799, 'epoch': 2.67} + 53%|█████▎ | 400/750 [41:30<34:42, 5.95s/it][INFO|trainer.py:3831] 2025-06-27 00:30:17,451 >> +***** Running Evaluation ***** +[INFO|trainer.py:3833] 2025-06-27 00:30:17,451 >> Num examples = 1000 +[INFO|trainer.py:3836] 2025-06-27 00:30:17,451 >> Batch size = 25 + + 0%| | 0/10 [00:00> Saving model checkpoint to ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400 +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d98ee-0476a86e6a6f1a56406e1487;c8f63932-7692-49dc-a585-c84f146d96b9) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +[INFO|tokenization_utils_base.py:2684] 2025-06-27 00:31:02,321 >> tokenizer config file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/tokenizer_config.json +[INFO|tokenization_utils_base.py:2693] 2025-06-27 00:31:02,321 >> Special tokens file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/special_tokens_map.json +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +[2025-06-27 00:31:04,807] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step400 is begin to save! +[2025-06-27 00:31:04,832] [INFO] [logging.py:107:log_dist] [Rank 0] Saving model checkpoint: ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/global_step400/mp_rank_00_model_states.pt +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d98f1-3de239f907cdb1d94d5c5249;5ba9ed54-ae74-4b07-b188-9efe6bc4b684) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d98f1-4994e1c043686ef74b93d7c3;fd3bf65f-88a1-43b8-892c-7c1cc80492da) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d98f1-7a95ed364dfc844d3064d52f;e3e3af4a-18e4-47d8-abdf-cfe0a5763a38) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d98f1-2861a5886db662a570c575c0;398e435c-532a-4da1-a166-cebd178350ff) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( + 53%|█████▎ | 401/750 [42:24<1:58:37, 20.39s/it] {'loss': 0.1893, 'grad_norm': 0.12802425026893616, 'learning_rate': 0.00011974619876931687, 'epoch': 2.67} + 53%|█████▎ | 401/750 [42:24<1:58:37, 20.39s/it] 54%|█████▎ | 402/750 [42:30<1:33:04, 16.05s/it] {'loss': 0.1748, 'grad_norm': 0.13873571157455444, 'learning_rate': 0.0001195971679725932, 'epoch': 2.68} + 54%|█████▎ | 402/750 [42:30<1:33:04, 16.05s/it] 54%|█████▎ | 403/750 [42:36<1:15:10, 13.00s/it] {'loss': 0.1738, 'grad_norm': 0.13294534385204315, 'learning_rate': 0.00011944869222574892, 'epoch': 2.69} + 54%|█████▎ | 403/750 [42:36<1:15:10, 13.00s/it] 54%|█████▍ | 404/750 [42:42<1:02:47, 10.89s/it] {'loss': 0.1721, 'grad_norm': 0.11917472630739212, 'learning_rate': 0.00011930076809193951, 'epoch': 2.69} + 54%|█████▍ | 404/750 [42:42<1:02:47, 10.89s/it] 54%|█████▍ | 405/750 [42:48<53:55, 9.38s/it] {'loss': 0.1784, 'grad_norm': 0.13650228083133698, 'learning_rate': 0.0001191533921640401, 'epoch': 2.7} + 54%|█████▍ | 405/750 [42:48<53:55, 9.38s/it] 54%|█████▍ | 406/750 [42:54<47:41, 8.32s/it] {'loss': 0.1786, 'grad_norm': 0.14056342840194702, 'learning_rate': 0.00011900656106431562, 'epoch': 2.71} + 54%|█████▍ | 406/750 [42:54<47:41, 8.32s/it] 54%|█████▍ | 407/750 [42:59<43:29, 7.61s/it] {'loss': 0.161, 'grad_norm': 0.13128921389579773, 'learning_rate': 0.00011886027144409578, 'epoch': 2.71} + 54%|█████▍ | 407/750 [42:59<43:29, 7.61s/it] 54%|█████▍ | 408/750 [43:05<40:41, 7.14s/it] {'loss': 0.1841, 'grad_norm': 0.1360347718000412, 'learning_rate': 0.00011871451998345427, 'epoch': 2.72} + 54%|█████▍ | 408/750 [43:06<40:41, 7.14s/it] 55%|█████▍ | 409/750 [43:11<38:23, 6.76s/it] {'loss': 0.1592, 'grad_norm': 0.13800328969955444, 'learning_rate': 0.00011856930339089229, 'epoch': 2.73} + 55%|█████▍ | 409/750 [43:11<38:23, 6.76s/it] 55%|█████▍ | 410/750 [43:18<37:17, 6.58s/it] {'loss': 0.1935, 'grad_norm': 0.1339549869298935, 'learning_rate': 0.00011842461840302649, 'epoch': 2.73} + 55%|█████▍ | 410/750 [43:18<37:17, 6.58s/it] 55%|█████▍ | 411/750 [43:24<36:21, 6.44s/it] {'loss': 0.1537, 'grad_norm': 0.12085722386837006, 'learning_rate': 0.00011828046178428064, 'epoch': 2.74} + 55%|█████▍ | 411/750 [43:24<36:21, 6.44s/it] 55%|█████▍ | 412/750 [43:30<35:29, 6.30s/it] {'loss': 0.1655, 'grad_norm': 0.1239975243806839, 'learning_rate': 0.00011813683032658212, 'epoch': 2.75} + 55%|█████▍ | 412/750 [43:30<35:29, 6.30s/it] 55%|█████▌ | 413/750 [43:36<34:48, 6.20s/it] {'loss': 0.1696, 'grad_norm': 0.1383960247039795, 'learning_rate': 0.0001179937208490617, 'epoch': 2.75} + 55%|█████▌ | 413/750 [43:36<34:48, 6.20s/it] 55%|█████▌ | 414/750 [43:42<34:20, 6.13s/it] {'loss': 0.1825, 'grad_norm': 0.13461098074913025, 'learning_rate': 0.00011785113019775794, 'epoch': 2.76} + 55%|█████▌ | 414/750 [43:42<34:20, 6.13s/it] 55%|█████▌ | 415/750 [43:48<33:59, 6.09s/it] {'loss': 0.1815, 'grad_norm': 0.13429348170757294, 'learning_rate': 0.00011770905524532507, 'epoch': 2.77} + 55%|█████▌ | 415/750 [43:48<33:59, 6.09s/it] 55%|█████▌ | 416/750 [43:54<33:50, 6.08s/it] {'loss': 0.177, 'grad_norm': 0.13084062933921814, 'learning_rate': 0.00011756749289074503, 'epoch': 2.77} + 55%|█████▌ | 416/750 [43:54<33:50, 6.08s/it] 56%|█████▌ | 417/750 [44:00<33:29, 6.04s/it] {'loss': 0.154, 'grad_norm': 0.1400175243616104, 'learning_rate': 0.00011742644005904313, 'epoch': 2.78} + 56%|█████▌ | 417/750 [44:00<33:29, 6.04s/it] 56%|█████▌ | 418/750 [44:05<33:12, 6.00s/it] {'loss': 0.1561, 'grad_norm': 0.13435116410255432, 'learning_rate': 0.00011728589370100743, 'epoch': 2.79} + 56%|█████▌ | 418/750 [44:05<33:12, 6.00s/it] 56%|█████▌ | 419/750 [44:11<33:02, 5.99s/it] {'loss': 0.1714, 'grad_norm': 0.14085592329502106, 'learning_rate': 0.00011714585079291212, 'epoch': 2.79} + 56%|█████▌ | 419/750 [44:11<33:02, 5.99s/it] 56%|█████▌ | 420/750 [44:17<32:56, 5.99s/it] {'loss': 0.1774, 'grad_norm': 0.13033795356750488, 'learning_rate': 0.00011700630833624395, 'epoch': 2.8} + 56%|█████▌ | 420/750 [44:17<32:56, 5.99s/it] 56%|█████▌ | 421/750 [44:23<32:46, 5.98s/it] {'loss': 0.1757, 'grad_norm': 0.13652119040489197, 'learning_rate': 0.00011686726335743291, 'epoch': 2.81} + 56%|█████▌ | 421/750 [44:23<32:46, 5.98s/it] 56%|█████▋ | 422/750 [44:29<32:50, 6.01s/it] {'loss': 0.185, 'grad_norm': 0.12395608425140381, 'learning_rate': 0.0001167287129075859, 'epoch': 2.81} + 56%|█████▋ | 422/750 [44:29<32:50, 6.01s/it] 56%|█████▋ | 423/750 [44:35<32:41, 6.00s/it] {'loss': 0.1743, 'grad_norm': 0.134079247713089, 'learning_rate': 0.00011659065406222409, 'epoch': 2.82} + 56%|█████▋ | 423/750 [44:35<32:41, 6.00s/it] 57%|█████▋ | 424/750 [44:41<32:32, 5.99s/it] {'loss': 0.1608, 'grad_norm': 0.1343086063861847, 'learning_rate': 0.00011645308392102366, 'epoch': 2.83} + 57%|█████▋ | 424/750 [44:41<32:32, 5.99s/it] 57%|█████▋ | 425/750 [44:47<32:13, 5.95s/it] {'loss': 0.1592, 'grad_norm': 0.12749235332012177, 'learning_rate': 0.00011631599960755992, 'epoch': 2.83} + 57%|█████▋ | 425/750 [44:47<32:13, 5.95s/it] 57%|█████▋ | 426/750 [44:53<32:12, 5.97s/it] {'loss': 0.17, 'grad_norm': 0.12273029237985611, 'learning_rate': 0.00011617939826905469, 'epoch': 2.84} + 57%|█████▋ | 426/750 [44:53<32:12, 5.97s/it] 57%|█████▋ | 427/750 [44:59<32:10, 5.98s/it] {'loss': 0.1658, 'grad_norm': 0.13161002099514008, 'learning_rate': 0.00011604327707612684, 'epoch': 2.85} + 57%|█████▋ | 427/750 [44:59<32:10, 5.98s/it] 57%|█████▋ | 428/750 [45:05<32:20, 6.03s/it] {'loss': 0.1589, 'grad_norm': 0.12151265889406204, 'learning_rate': 0.00011590763322254638, 'epoch': 2.85} + 57%|█████▋ | 428/750 [45:05<32:20, 6.03s/it] 57%|█████▋ | 429/750 [45:11<32:01, 5.99s/it] {'loss': 0.1783, 'grad_norm': 0.14203910529613495, 'learning_rate': 0.00011577246392499127, 'epoch': 2.86} + 57%|█████▋ | 429/750 [45:11<32:01, 5.99s/it] 57%|█████▋ | 430/750 [45:17<32:06, 6.02s/it] {'loss': 0.1545, 'grad_norm': 0.13163644075393677, 'learning_rate': 0.00011563776642280763, 'epoch': 2.87} + 57%|█████▋ | 430/750 [45:17<32:06, 6.02s/it] 57%|█████▋ | 431/750 [45:23<31:56, 6.01s/it] {'loss': 0.1663, 'grad_norm': 0.13118694722652435, 'learning_rate': 0.00011550353797777291, 'epoch': 2.87} + 57%|█████▋ | 431/750 [45:23<31:56, 6.01s/it] 58%|█████▊ | 432/750 [45:29<31:43, 5.99s/it] {'loss': 0.1593, 'grad_norm': 0.12807047367095947, 'learning_rate': 0.00011536977587386216, 'epoch': 2.88} + 58%|█████▊ | 432/750 [45:29<31:43, 5.99s/it] 58%|█████▊ | 433/750 [45:35<31:38, 5.99s/it] {'loss': 0.1518, 'grad_norm': 0.12787529826164246, 'learning_rate': 0.00011523647741701704, 'epoch': 2.89} + 58%|█████▊ | 433/750 [45:35<31:38, 5.99s/it] 58%|█████▊ | 434/750 [45:41<31:31, 5.98s/it] {'loss': 0.178, 'grad_norm': 0.13488884270191193, 'learning_rate': 0.00011510363993491814, 'epoch': 2.89} + 58%|█████▊ | 434/750 [45:41<31:31, 5.98s/it] 58%|█████▊ | 435/750 [45:47<31:25, 5.99s/it] {'loss': 0.178, 'grad_norm': 0.13938601315021515, 'learning_rate': 0.00011497126077675976, 'epoch': 2.9} + 58%|█████▊ | 435/750 [45:47<31:25, 5.99s/it] 58%|█████▊ | 436/750 [45:53<31:11, 5.96s/it] {'loss': 0.1897, 'grad_norm': 0.1396407037973404, 'learning_rate': 0.00011483933731302795, 'epoch': 2.91} + 58%|█████▊ | 436/750 [45:53<31:11, 5.96s/it] 58%|█████▊ | 437/750 [45:59<31:01, 5.95s/it] {'loss': 0.1697, 'grad_norm': 0.12715566158294678, 'learning_rate': 0.00011470786693528087, 'epoch': 2.91} + 58%|█████▊ | 437/750 [45:59<31:01, 5.95s/it] 58%|█████▊ | 438/750 [46:05<30:48, 5.92s/it] {'loss': 0.1879, 'grad_norm': 0.13826040923595428, 'learning_rate': 0.00011457684705593247, 'epoch': 2.92} + 58%|█████▊ | 438/750 [46:05<30:48, 5.92s/it] 59%|█████▊ | 439/750 [46:11<30:44, 5.93s/it] {'loss': 0.1869, 'grad_norm': 0.13141952455043793, 'learning_rate': 0.00011444627510803829, 'epoch': 2.93} + 59%|█████▊ | 439/750 [46:11<30:44, 5.93s/it] 59%|█████▊ | 440/750 [46:17<30:33, 5.91s/it] {'loss': 0.1625, 'grad_norm': 0.1258581280708313, 'learning_rate': 0.00011431614854508447, 'epoch': 2.93} + 59%|█████▊ | 440/750 [46:17<30:33, 5.91s/it] 59%|█████▉ | 441/750 [46:23<30:33, 5.93s/it] {'loss': 0.1499, 'grad_norm': 0.12165071070194244, 'learning_rate': 0.00011418646484077904, 'epoch': 2.94} + 59%|█████▉ | 441/750 [46:23<30:33, 5.93s/it] 59%|█████▉ | 442/750 [46:29<30:30, 5.94s/it] {'loss': 0.1591, 'grad_norm': 0.1242181733250618, 'learning_rate': 0.00011405722148884592, 'epoch': 2.95} + 59%|█████▉ | 442/750 [46:29<30:30, 5.94s/it] 59%|█████▉ | 443/750 [46:35<30:22, 5.94s/it] {'loss': 0.1737, 'grad_norm': 0.13902366161346436, 'learning_rate': 0.00011392841600282156, 'epoch': 2.95} + 59%|█████▉ | 443/750 [46:35<30:22, 5.94s/it] 59%|█████▉ | 444/750 [46:41<30:12, 5.92s/it] {'loss': 0.1972, 'grad_norm': 0.13813108205795288, 'learning_rate': 0.00011380004591585387, 'epoch': 2.96} + 59%|█████▉ | 444/750 [46:41<30:12, 5.92s/it] 59%|█████▉ | 445/750 [46:46<30:05, 5.92s/it] {'loss': 0.1746, 'grad_norm': 0.13776984810829163, 'learning_rate': 0.00011367210878050402, 'epoch': 2.97} + 59%|█████▉ | 445/750 [46:46<30:05, 5.92s/it] 59%|█████▉ | 446/750 [46:52<29:58, 5.91s/it] {'loss': 0.1691, 'grad_norm': 0.13031359016895294, 'learning_rate': 0.00011354460216855028, 'epoch': 2.97} + 59%|█████▉ | 446/750 [46:52<29:58, 5.91s/it] 60%|█████▉ | 447/750 [46:58<30:00, 5.94s/it] {'loss': 0.1791, 'grad_norm': 0.1299438327550888, 'learning_rate': 0.00011341752367079442, 'epoch': 2.98} + 60%|█████▉ | 447/750 [46:58<30:00, 5.94s/it] 60%|█████▉ | 448/750 [47:04<30:10, 6.00s/it] {'loss': 0.175, 'grad_norm': 0.13401517271995544, 'learning_rate': 0.00011329087089687071, 'epoch': 2.99} + 60%|█████▉ | 448/750 [47:04<30:10, 6.00s/it] 60%|█████▉ | 449/750 [47:10<30:07, 6.00s/it] {'loss': 0.1818, 'grad_norm': 0.14243927597999573, 'learning_rate': 0.00011316464147505693, 'epoch': 2.99} + 60%|█████▉ | 449/750 [47:10<30:07, 6.00s/it] 60%|██████ | 450/750 [47:18<32:57, 6.59s/it] {'loss': 0.1652, 'grad_norm': 0.1279725432395935, 'learning_rate': 0.0001130388330520878, 'epoch': 3.0} + 60%|██████ | 450/750 [47:18<32:57, 6.59s/it] 60%|██████ | 451/750 [47:33<44:49, 8.99s/it] {'loss': 0.1655, 'grad_norm': 0.1305849254131317, 'learning_rate': 0.00011291344329297086, 'epoch': 3.01} + 60%|██████ | 451/750 [47:33<44:49, 8.99s/it] 60%|██████ | 452/750 [47:39<40:07, 8.08s/it] {'loss': 0.1494, 'grad_norm': 0.12583260238170624, 'learning_rate': 0.0001127884698808043, 'epoch': 3.01} + 60%|██████ | 452/750 [47:39<40:07, 8.08s/it] 60%|██████ | 453/750 [47:45<36:50, 7.44s/it] {'loss': 0.1321, 'grad_norm': 0.12696702778339386, 'learning_rate': 0.00011266391051659725, 'epoch': 3.02} + 60%|██████ | 453/750 [47:45<36:50, 7.44s/it] 61%|██████ | 454/750 [47:51<34:32, 7.00s/it] {'loss': 0.1292, 'grad_norm': 0.12244350463151932, 'learning_rate': 0.00011253976291909212, 'epoch': 3.03} + 61%|██████ | 454/750 [47:51<34:32, 7.00s/it] 61%|██████ | 455/750 [47:57<32:52, 6.69s/it] {'loss': 0.1409, 'grad_norm': 0.1371058225631714, 'learning_rate': 0.00011241602482458911, 'epoch': 3.03} + 61%|██████ | 455/750 [47:57<32:52, 6.69s/it] 61%|██████ | 456/750 [48:03<31:36, 6.45s/it] {'loss': 0.1306, 'grad_norm': 0.15259167551994324, 'learning_rate': 0.00011229269398677288, 'epoch': 3.04} + 61%|██████ | 456/750 [48:03<31:36, 6.45s/it] 61%|██████ | 457/750 [48:09<30:38, 6.28s/it] {'loss': 0.1329, 'grad_norm': 0.15388865768909454, 'learning_rate': 0.00011216976817654119, 'epoch': 3.05} + 61%|██████ | 457/750 [48:09<30:38, 6.28s/it] 61%|██████ | 458/750 [48:15<30:05, 6.18s/it] {'loss': 0.14, 'grad_norm': 0.15698152780532837, 'learning_rate': 0.00011204724518183582, 'epoch': 3.05} + 61%|██████ | 458/750 [48:15<30:05, 6.18s/it] 61%|██████ | 459/750 [48:21<29:43, 6.13s/it] {'loss': 0.147, 'grad_norm': 0.15454934537410736, 'learning_rate': 0.00011192512280747522, 'epoch': 3.06} + 61%|██████ | 459/750 [48:21<29:43, 6.13s/it] 61%|██████▏ | 460/750 [48:27<29:27, 6.10s/it] {'loss': 0.1612, 'grad_norm': 0.1518414169549942, 'learning_rate': 0.00011180339887498949, 'epoch': 3.07} + 61%|██████▏ | 460/750 [48:27<29:27, 6.10s/it] 61%|██████▏ | 461/750 [48:33<29:06, 6.04s/it] {'loss': 0.1432, 'grad_norm': 0.13643917441368103, 'learning_rate': 0.000111682071222457, 'epoch': 3.07} + 61%|██████▏ | 461/750 [48:33<29:06, 6.04s/it] 62%|██████▏ | 462/750 [48:38<28:46, 6.00s/it] {'loss': 0.1565, 'grad_norm': 0.14681023359298706, 'learning_rate': 0.00011156113770434329, 'epoch': 3.08} + 62%|██████▏ | 462/750 [48:38<28:46, 6.00s/it] 62%|██████▏ | 463/750 [48:44<28:38, 5.99s/it] {'loss': 0.1387, 'grad_norm': 0.13807366788387299, 'learning_rate': 0.00011144059619134151, 'epoch': 3.09} + 62%|██████▏ | 463/750 [48:44<28:38, 5.99s/it] 62%|██████▏ | 464/750 [48:50<28:30, 5.98s/it] {'loss': 0.1257, 'grad_norm': 0.13597892224788666, 'learning_rate': 0.00011132044457021517, 'epoch': 3.09} + 62%|██████▏ | 464/750 [48:50<28:30, 5.98s/it] 62%|██████▏ | 465/750 [48:56<28:13, 5.94s/it] {'loss': 0.1345, 'grad_norm': 0.14555016160011292, 'learning_rate': 0.00011120068074364232, 'epoch': 3.1} + 62%|██████▏ | 465/750 [48:56<28:13, 5.94s/it] 62%|██████▏ | 466/750 [49:02<28:20, 5.99s/it] {'loss': 0.132, 'grad_norm': 0.15187765657901764, 'learning_rate': 0.000111081302630062, 'epoch': 3.11} + 62%|██████▏ | 466/750 [49:02<28:20, 5.99s/it] 62%|██████▏ | 467/750 [49:08<28:12, 5.98s/it] {'loss': 0.1325, 'grad_norm': 0.14446990191936493, 'learning_rate': 0.0001109623081635221, 'epoch': 3.11} + 62%|██████▏ | 467/750 [49:08<28:12, 5.98s/it] 62%|██████▏ | 468/750 [49:14<28:16, 6.02s/it] {'loss': 0.1582, 'grad_norm': 0.16507649421691895, 'learning_rate': 0.00011084369529352937, 'epoch': 3.12} + 62%|██████▏ | 468/750 [49:14<28:16, 6.02s/it] 63%|██████▎ | 469/750 [49:20<27:57, 5.97s/it] {'loss': 0.1382, 'grad_norm': 0.16875247657299042, 'learning_rate': 0.00011072546198490086, 'epoch': 3.13} + 63%|██████▎ | 469/750 [49:20<27:57, 5.97s/it] 63%|██████▎ | 470/750 [49:26<27:48, 5.96s/it] {'loss': 0.1348, 'grad_norm': 0.1724245548248291, 'learning_rate': 0.00011060760621761728, 'epoch': 3.13} + 63%|██████▎ | 470/750 [49:26<27:48, 5.96s/it] 63%|██████▎ | 471/750 [49:32<27:43, 5.96s/it] {'loss': 0.1568, 'grad_norm': 0.15607628226280212, 'learning_rate': 0.00011049012598667817, 'epoch': 3.14} + 63%|██████▎ | 471/750 [49:32<27:43, 5.96s/it] 63%|██████▎ | 472/750 [49:38<27:32, 5.94s/it] {'loss': 0.1417, 'grad_norm': 0.1528671830892563, 'learning_rate': 0.00011037301930195852, 'epoch': 3.15} + 63%|██████▎ | 472/750 [49:38<27:32, 5.94s/it] 63%|██████▎ | 473/750 [49:44<27:17, 5.91s/it] {'loss': 0.1316, 'grad_norm': 0.1656087040901184, 'learning_rate': 0.00011025628418806717, 'epoch': 3.15} + 63%|██████▎ | 473/750 [49:44<27:17, 5.91s/it] 63%|██████▎ | 474/750 [49:50<27:13, 5.92s/it] {'loss': 0.1432, 'grad_norm': 0.15559254586696625, 'learning_rate': 0.00011013991868420692, 'epoch': 3.16} + 63%|██████▎ | 474/750 [49:50<27:13, 5.92s/it] 63%|██████▎ | 475/750 [49:56<27:08, 5.92s/it] {'loss': 0.1329, 'grad_norm': 0.1453288197517395, 'learning_rate': 0.00011002392084403617, 'epoch': 3.17} + 63%|██████▎ | 475/750 [49:56<27:08, 5.92s/it] 63%|██████▎ | 476/750 [50:02<26:56, 5.90s/it] {'loss': 0.1489, 'grad_norm': 0.1557648926973343, 'learning_rate': 0.00010990828873553218, 'epoch': 3.17} + 63%|██████▎ | 476/750 [50:02<26:56, 5.90s/it] 64%|██████▎ | 477/750 [50:07<26:49, 5.90s/it] {'loss': 0.1489, 'grad_norm': 0.15618963539600372, 'learning_rate': 0.00010979302044085591, 'epoch': 3.18} + 64%|██████▎ | 477/750 [50:07<26:49, 5.90s/it] 64%|██████▎ | 478/750 [50:13<26:42, 5.89s/it] {'loss': 0.1365, 'grad_norm': 0.15531198680400848, 'learning_rate': 0.00010967811405621847, 'epoch': 3.19} + 64%|██████▎ | 478/750 [50:13<26:42, 5.89s/it] 64%|██████▍ | 479/750 [50:19<26:43, 5.92s/it] {'loss': 0.1347, 'grad_norm': 0.14251747727394104, 'learning_rate': 0.000109563567691749, 'epoch': 3.19} + 64%|██████▍ | 479/750 [50:19<26:43, 5.92s/it] 64%|██████▍ | 480/750 [50:25<26:43, 5.94s/it] {'loss': 0.1509, 'grad_norm': 0.15274056792259216, 'learning_rate': 0.00010944937947136412, 'epoch': 3.2} + 64%|██████▍ | 480/750 [50:25<26:43, 5.94s/it] 64%|██████▍ | 481/750 [50:31<26:38, 5.94s/it] {'loss': 0.1289, 'grad_norm': 0.14671817421913147, 'learning_rate': 0.00010933554753263895, 'epoch': 3.21} + 64%|██████▍ | 481/750 [50:31<26:38, 5.94s/it] 64%|██████▍ | 482/750 [50:37<26:28, 5.93s/it] {'loss': 0.1385, 'grad_norm': 0.16423137485980988, 'learning_rate': 0.00010922207002667943, 'epoch': 3.21} + 64%|██████▍ | 482/750 [50:37<26:28, 5.93s/it] 64%|██████▍ | 483/750 [50:43<26:20, 5.92s/it] {'loss': 0.135, 'grad_norm': 0.1560370773077011, 'learning_rate': 0.0001091089451179962, 'epoch': 3.22} + 64%|██████▍ | 483/750 [50:43<26:20, 5.92s/it] 65%|██████▍ | 484/750 [50:49<26:16, 5.93s/it] {'loss': 0.1536, 'grad_norm': 0.16102741658687592, 'learning_rate': 0.00010899617098437999, 'epoch': 3.23} + 65%|██████▍ | 484/750 [50:49<26:16, 5.93s/it] 65%|██████▍ | 485/750 [50:55<26:12, 5.93s/it] {'loss': 0.1496, 'grad_norm': 0.1637362390756607, 'learning_rate': 0.00010888374581677828, 'epoch': 3.23} + 65%|██████▍ | 485/750 [50:55<26:12, 5.93s/it] 65%|██████▍ | 486/750 [51:01<26:05, 5.93s/it] {'loss': 0.1405, 'grad_norm': 0.152584508061409, 'learning_rate': 0.00010877166781917343, 'epoch': 3.24} + 65%|██████▍ | 486/750 [51:01<26:05, 5.93s/it] 65%|██████▍ | 487/750 [51:07<26:11, 5.98s/it] {'loss': 0.1574, 'grad_norm': 0.17232313752174377, 'learning_rate': 0.00010865993520846211, 'epoch': 3.25} + 65%|██████▍ | 487/750 [51:07<26:11, 5.98s/it] 65%|██████▌ | 488/750 [51:13<25:58, 5.95s/it] {'loss': 0.142, 'grad_norm': 0.1526651233434677, 'learning_rate': 0.0001085485462143363, 'epoch': 3.25} + 65%|██████▌ | 488/750 [51:13<25:58, 5.95s/it] 65%|██████▌ | 489/750 [51:19<25:47, 5.93s/it] {'loss': 0.1349, 'grad_norm': 0.15860426425933838, 'learning_rate': 0.00010843749907916526, 'epoch': 3.26} + 65%|██████▌ | 489/750 [51:19<25:47, 5.93s/it] 65%|██████▌ | 490/750 [51:25<25:32, 5.89s/it] {'loss': 0.1484, 'grad_norm': 0.16029545664787292, 'learning_rate': 0.0001083267920578793, 'epoch': 3.27} + 65%|██████▌ | 490/750 [51:25<25:32, 5.89s/it] 65%|██████▌ | 491/750 [51:30<25:31, 5.91s/it] {'loss': 0.1388, 'grad_norm': 0.15227065980434418, 'learning_rate': 0.00010821642341785439, 'epoch': 3.27} + 65%|██████▌ | 491/750 [51:30<25:31, 5.91s/it] 66%|██████▌ | 492/750 [51:36<25:26, 5.92s/it] {'loss': 0.156, 'grad_norm': 0.1688096970319748, 'learning_rate': 0.00010810639143879838, 'epoch': 3.28} + 66%|██████▌ | 492/750 [51:36<25:26, 5.92s/it] 66%|██████▌ | 493/750 [51:42<25:17, 5.90s/it] {'loss': 0.1513, 'grad_norm': 0.15941110253334045, 'learning_rate': 0.00010799669441263838, 'epoch': 3.29} + 66%|██████▌ | 493/750 [51:42<25:17, 5.90s/it] 66%|██████▌ | 494/750 [51:48<25:07, 5.89s/it] {'loss': 0.1483, 'grad_norm': 0.16796240210533142, 'learning_rate': 0.0001078873306434093, 'epoch': 3.29} + 66%|██████▌ | 494/750 [51:48<25:07, 5.89s/it] 66%|██████▌ | 495/750 [51:54<25:06, 5.91s/it] {'loss': 0.1561, 'grad_norm': 0.17123353481292725, 'learning_rate': 0.00010777829844714388, 'epoch': 3.3} + 66%|██████▌ | 495/750 [51:54<25:06, 5.91s/it] 66%|██████▌ | 496/750 [52:00<25:06, 5.93s/it] {'loss': 0.134, 'grad_norm': 0.15404053032398224, 'learning_rate': 0.0001076695961517636, 'epoch': 3.31} + 66%|██████▌ | 496/750 [52:00<25:06, 5.93s/it] 66%|██████▋ | 497/750 [52:06<25:02, 5.94s/it] {'loss': 0.1471, 'grad_norm': 0.1669268012046814, 'learning_rate': 0.0001075612220969712, 'epoch': 3.31} + 66%|██████▋ | 497/750 [52:06<25:02, 5.94s/it] 66%|██████▋ | 498/750 [52:12<24:49, 5.91s/it] {'loss': 0.1379, 'grad_norm': 0.1644563376903534, 'learning_rate': 0.00010745317463414389, 'epoch': 3.32} + 66%|██████▋ | 498/750 [52:12<24:49, 5.91s/it] 67%|██████▋ | 499/750 [52:18<24:48, 5.93s/it] {'loss': 0.1357, 'grad_norm': 0.1556372046470642, 'learning_rate': 0.00010734545212622828, 'epoch': 3.33} + 67%|██████▋ | 499/750 [52:18<24:48, 5.93s/it] 67%|██████▋ | 500/750 [52:24<24:37, 5.91s/it] {'loss': 0.1669, 'grad_norm': 0.17315615713596344, 'learning_rate': 0.00010723805294763608, 'epoch': 3.33} + 67%|██████▋ | 500/750 [52:24<24:37, 5.91s/it] 67%|██████▋ | 501/750 [52:30<24:32, 5.91s/it] {'loss': 0.161, 'grad_norm': 0.1610376238822937, 'learning_rate': 0.00010713097548414107, 'epoch': 3.34} + 67%|██████▋ | 501/750 [52:30<24:32, 5.91s/it] 67%|██████▋ | 502/750 [52:36<24:28, 5.92s/it] {'loss': 0.1498, 'grad_norm': 0.16572323441505432, 'learning_rate': 0.00010702421813277718, 'epoch': 3.35} + 67%|██████▋ | 502/750 [52:36<24:28, 5.92s/it] 67%|██████▋ | 503/750 [52:42<24:34, 5.97s/it] {'loss': 0.1544, 'grad_norm': 0.16042861342430115, 'learning_rate': 0.0001069177793017378, 'epoch': 3.35} + 67%|██████▋ | 503/750 [52:42<24:34, 5.97s/it] 67%|██████▋ | 504/750 [52:48<24:26, 5.96s/it] {'loss': 0.1543, 'grad_norm': 0.1622343510389328, 'learning_rate': 0.00010681165741027595, 'epoch': 3.36} + 67%|██████▋ | 504/750 [52:48<24:26, 5.96s/it] 67%|██████▋ | 505/750 [52:54<24:19, 5.96s/it] {'loss': 0.1433, 'grad_norm': 0.15245985984802246, 'learning_rate': 0.00010670585088860586, 'epoch': 3.37} + 67%|██████▋ | 505/750 [52:54<24:19, 5.96s/it] 67%|██████▋ | 506/750 [53:00<24:27, 6.01s/it] {'loss': 0.1526, 'grad_norm': 0.15931057929992676, 'learning_rate': 0.00010660035817780521, 'epoch': 3.37} + 67%|██████▋ | 506/750 [53:00<24:27, 6.01s/it] 68%|██████▊ | 507/750 [53:06<24:19, 6.01s/it] {'loss': 0.1361, 'grad_norm': 0.15014851093292236, 'learning_rate': 0.00010649517772971891, 'epoch': 3.38} + 68%|██████▊ | 507/750 [53:06<24:19, 6.01s/it] 68%|██████▊ | 508/750 [53:12<24:04, 5.97s/it] {'loss': 0.1247, 'grad_norm': 0.16248862445354462, 'learning_rate': 0.00010639030800686347, 'epoch': 3.39} + 68%|██████▊ | 508/750 [53:12<24:04, 5.97s/it] 68%|██████▊ | 509/750 [53:18<23:58, 5.97s/it] {'loss': 0.1353, 'grad_norm': 0.1534966379404068, 'learning_rate': 0.0001062857474823327, 'epoch': 3.39} + 68%|██████▊ | 509/750 [53:18<23:58, 5.97s/it] 68%|██████▊ | 510/750 [53:23<23:51, 5.97s/it] {'loss': 0.1464, 'grad_norm': 0.15869368612766266, 'learning_rate': 0.00010618149463970438, 'epoch': 3.4} + 68%|██████▊ | 510/750 [53:24<23:51, 5.97s/it] 68%|██████▊ | 511/750 [53:29<23:44, 5.96s/it] {'loss': 0.136, 'grad_norm': 0.14489006996154785, 'learning_rate': 0.00010607754797294777, 'epoch': 3.41} + 68%|██████▊ | 511/750 [53:29<23:44, 5.96s/it] 68%|██████▊ | 512/750 [53:35<23:37, 5.95s/it] {'loss': 0.1537, 'grad_norm': 0.15689688920974731, 'learning_rate': 0.00010597390598633231, 'epoch': 3.41} + 68%|██████▊ | 512/750 [53:35<23:37, 5.95s/it] 68%|██████▊ | 513/750 [53:41<23:29, 5.95s/it] {'loss': 0.1497, 'grad_norm': 0.17782750725746155, 'learning_rate': 0.00010587056719433728, 'epoch': 3.42} + 68%|██████▊ | 513/750 [53:41<23:29, 5.95s/it] 69%|██████▊ | 514/750 [53:47<23:21, 5.94s/it] {'loss': 0.1424, 'grad_norm': 0.15875929594039917, 'learning_rate': 0.00010576753012156217, 'epoch': 3.43} + 69%|██████▊ | 514/750 [53:47<23:21, 5.94s/it] 69%|██████▊ | 515/750 [53:53<23:06, 5.90s/it] {'loss': 0.1523, 'grad_norm': 0.1707359254360199, 'learning_rate': 0.0001056647933026385, 'epoch': 3.43} + 69%|██████▊ | 515/750 [53:53<23:06, 5.90s/it] 69%|██████▉ | 516/750 [53:59<22:57, 5.89s/it] {'loss': 0.1594, 'grad_norm': 0.1630873680114746, 'learning_rate': 0.00010556235528214203, 'epoch': 3.44} + 69%|██████▉ | 516/750 [53:59<22:57, 5.89s/it] 69%|██████▉ | 517/750 [54:05<22:56, 5.91s/it] {'loss': 0.1418, 'grad_norm': 0.15407094359397888, 'learning_rate': 0.00010546021461450624, 'epoch': 3.45} + 69%|██████▉ | 517/750 [54:05<22:56, 5.91s/it] 69%|██████▉ | 518/750 [54:11<22:46, 5.89s/it] {'loss': 0.1531, 'grad_norm': 0.1630760133266449, 'learning_rate': 0.00010535836986393678, 'epoch': 3.45} + 69%|██████▉ | 518/750 [54:11<22:46, 5.89s/it] 69%|██████▉ | 519/750 [54:17<22:40, 5.89s/it] {'loss': 0.1638, 'grad_norm': 0.17024356126785278, 'learning_rate': 0.00010525681960432646, 'epoch': 3.46} + 69%|██████▉ | 519/750 [54:17<22:40, 5.89s/it] 69%|██████▉ | 520/750 [54:23<22:38, 5.91s/it] {'loss': 0.1383, 'grad_norm': 0.158006951212883, 'learning_rate': 0.00010515556241917166, 'epoch': 3.47} + 69%|██████▉ | 520/750 [54:23<22:38, 5.91s/it] 69%|██████▉ | 521/750 [54:28<22:30, 5.90s/it] {'loss': 0.1584, 'grad_norm': 0.15718567371368408, 'learning_rate': 0.00010505459690148908, 'epoch': 3.47} + 69%|██████▉ | 521/750 [54:28<22:30, 5.90s/it] 70%|██████▉ | 522/750 [54:34<22:26, 5.91s/it] {'loss': 0.1508, 'grad_norm': 0.16239719092845917, 'learning_rate': 0.00010495392165373376, 'epoch': 3.48} + 70%|██████▉ | 522/750 [54:34<22:26, 5.91s/it] 70%|██████▉ | 523/750 [54:40<22:32, 5.96s/it] {'loss': 0.1432, 'grad_norm': 0.1703788787126541, 'learning_rate': 0.00010485353528771787, 'epoch': 3.49} + 70%|██████▉ | 523/750 [54:40<22:32, 5.96s/it] 70%|██████▉ | 524/750 [54:46<22:22, 5.94s/it] {'loss': 0.1347, 'grad_norm': 0.1606765240430832, 'learning_rate': 0.00010475343642453021, 'epoch': 3.49} + 70%|██████▉ | 524/750 [54:46<22:22, 5.94s/it] 70%|███████ | 525/750 [54:52<22:24, 5.97s/it] {'loss': 0.1377, 'grad_norm': 0.1564657986164093, 'learning_rate': 0.00010465362369445672, 'epoch': 3.5} + 70%|███████ | 525/750 [54:52<22:24, 5.97s/it] 70%|███████ | 526/750 [54:58<22:19, 5.98s/it] {'loss': 0.1491, 'grad_norm': 0.15935762226581573, 'learning_rate': 0.00010455409573690175, 'epoch': 3.51} + 70%|███████ | 526/750 [54:58<22:19, 5.98s/it] 70%|███████ | 527/750 [55:04<22:13, 5.98s/it] {'loss': 0.1531, 'grad_norm': 0.16481848061084747, 'learning_rate': 0.00010445485120031016, 'epoch': 3.51} + 70%|███████ | 527/750 [55:04<22:13, 5.98s/it] 70%|███████ | 528/750 [55:10<22:01, 5.95s/it] {'loss': 0.1392, 'grad_norm': 0.15417121350765228, 'learning_rate': 0.00010435588874209024, 'epoch': 3.52} + 70%|███████ | 528/750 [55:10<22:01, 5.95s/it] 71%|███████ | 529/750 [55:16<21:56, 5.96s/it] {'loss': 0.1455, 'grad_norm': 0.15800175070762634, 'learning_rate': 0.00010425720702853738, 'epoch': 3.53} + 71%|███████ | 529/750 [55:16<21:56, 5.96s/it] 71%|███████ | 530/750 [55:22<21:54, 5.98s/it] {'loss': 0.1533, 'grad_norm': 0.1587381511926651, 'learning_rate': 0.00010415880473475868, 'epoch': 3.53} + 71%|███████ | 530/750 [55:22<21:54, 5.98s/it] 71%|███████ | 531/750 [55:28<21:46, 5.97s/it] {'loss': 0.1462, 'grad_norm': 0.15233056247234344, 'learning_rate': 0.0001040606805445981, 'epoch': 3.54} + 71%|███████ | 531/750 [55:28<21:46, 5.97s/it] 71%|███████ | 532/750 [55:34<21:30, 5.92s/it] {'loss': 0.1415, 'grad_norm': 0.1552983820438385, 'learning_rate': 0.00010396283315056266, 'epoch': 3.55} + 71%|███████ | 532/750 [55:34<21:30, 5.92s/it] 71%|███████ | 533/750 [55:40<21:26, 5.93s/it] {'loss': 0.1435, 'grad_norm': 0.14926421642303467, 'learning_rate': 0.00010386526125374917, 'epoch': 3.55} + 71%|███████ | 533/750 [55:40<21:26, 5.93s/it] 71%|███████ | 534/750 [55:46<21:16, 5.91s/it] {'loss': 0.1498, 'grad_norm': 0.1699218600988388, 'learning_rate': 0.00010376796356377186, 'epoch': 3.56} + 71%|███████ | 534/750 [55:46<21:16, 5.91s/it] 71%|███████▏ | 535/750 [55:52<21:13, 5.92s/it] {'loss': 0.1539, 'grad_norm': 0.16543714702129364, 'learning_rate': 0.00010367093879869068, 'epoch': 3.57} + 71%|███████▏ | 535/750 [55:52<21:13, 5.92s/it] 71%|███████▏ | 536/750 [55:58<21:09, 5.93s/it] {'loss': 0.1517, 'grad_norm': 0.15510369837284088, 'learning_rate': 0.00010357418568494037, 'epoch': 3.57} + 71%|███████▏ | 536/750 [55:58<21:09, 5.93s/it] 72%|███████▏ | 537/750 [56:04<20:57, 5.91s/it] {'loss': 0.1444, 'grad_norm': 0.1694919764995575, 'learning_rate': 0.00010347770295726023, 'epoch': 3.58} + 72%|███████▏ | 537/750 [56:04<20:57, 5.91s/it] 72%|███████▏ | 538/750 [56:09<20:54, 5.92s/it] {'loss': 0.1496, 'grad_norm': 0.1562710702419281, 'learning_rate': 0.00010338148935862473, 'epoch': 3.59} + 72%|███████▏ | 538/750 [56:09<20:54, 5.92s/it] 72%|███████▏ | 539/750 [56:15<20:50, 5.92s/it] {'loss': 0.1501, 'grad_norm': 0.16131797432899475, 'learning_rate': 0.00010328554364017445, 'epoch': 3.59} + 72%|███████▏ | 539/750 [56:15<20:50, 5.92s/it] 72%|███████▏ | 540/750 [56:21<20:47, 5.94s/it] {'loss': 0.1477, 'grad_norm': 0.14995840191841125, 'learning_rate': 0.00010318986456114837, 'epoch': 3.6} + 72%|███████▏ | 540/750 [56:21<20:47, 5.94s/it] 72%|███████▏ | 541/750 [56:28<20:55, 6.01s/it] {'loss': 0.1444, 'grad_norm': 0.15870153903961182, 'learning_rate': 0.00010309445088881625, 'epoch': 3.61} + 72%|███████▏ | 541/750 [56:28<20:55, 6.01s/it] 72%|███████▏ | 542/750 [56:33<20:39, 5.96s/it] {'loss': 0.15, 'grad_norm': 0.16403728723526, 'learning_rate': 0.00010299930139841185, 'epoch': 3.61} + 72%|███████▏ | 542/750 [56:33<20:39, 5.96s/it] 72%|███████▏ | 543/750 [56:39<20:41, 6.00s/it] {'loss': 0.1457, 'grad_norm': 0.14916802942752838, 'learning_rate': 0.00010290441487306716, 'epoch': 3.62} + 72%|███████▏ | 543/750 [56:39<20:41, 6.00s/it] 73%|███████▎ | 544/750 [56:45<20:11, 5.88s/it] {'loss': 0.1304, 'grad_norm': 0.15775884687900543, 'learning_rate': 0.00010280979010374679, 'epoch': 3.63} + 73%|███████▎ | 544/750 [56:45<20:11, 5.88s/it] 73%|███████▎ | 545/750 [56:51<20:03, 5.87s/it] {'loss': 0.1373, 'grad_norm': 0.17496716976165771, 'learning_rate': 0.00010271542588918341, 'epoch': 3.63} + 73%|███████▎ | 545/750 [56:51<20:03, 5.87s/it] 73%|███████▎ | 546/750 [56:57<20:07, 5.92s/it] {'loss': 0.1407, 'grad_norm': 0.15421578288078308, 'learning_rate': 0.00010262132103581367, 'epoch': 3.64} + 73%|███████▎ | 546/750 [56:57<20:07, 5.92s/it] 73%|███████▎ | 547/750 [57:03<20:04, 5.93s/it] {'loss': 0.1456, 'grad_norm': 0.15761473774909973, 'learning_rate': 0.0001025274743577147, 'epoch': 3.65} + 73%|███████▎ | 547/750 [57:03<20:04, 5.93s/it] 73%|███████▎ | 548/750 [57:09<19:58, 5.93s/it] {'loss': 0.1504, 'grad_norm': 0.16617563366889954, 'learning_rate': 0.00010243388467654152, 'epoch': 3.65} + 73%|███████▎ | 548/750 [57:09<19:58, 5.93s/it] 73%|███████▎ | 549/750 [57:15<19:54, 5.95s/it] {'loss': 0.1492, 'grad_norm': 0.1630965769290924, 'learning_rate': 0.0001023405508214647, 'epoch': 3.66} + 73%|███████▎ | 549/750 [57:15<19:54, 5.95s/it] 73%|███████▎ | 550/750 [57:21<19:44, 5.92s/it] {'loss': 0.1787, 'grad_norm': 0.18093156814575195, 'learning_rate': 0.00010224747162910903, 'epoch': 3.67} + 73%|███████▎ | 550/750 [57:21<19:44, 5.92s/it] 73%|███████▎ | 551/750 [57:27<19:38, 5.92s/it] {'loss': 0.1535, 'grad_norm': 0.1675328016281128, 'learning_rate': 0.00010215464594349242, 'epoch': 3.67} + 73%|███████▎ | 551/750 [57:27<19:38, 5.92s/it] 74%|███████▎ | 552/750 [57:33<19:29, 5.91s/it] {'loss': 0.1493, 'grad_norm': 0.1581062376499176, 'learning_rate': 0.00010206207261596577, 'epoch': 3.68} + 74%|███████▎ | 552/750 [57:33<19:29, 5.91s/it] 74%|███████▎ | 553/750 [57:38<19:23, 5.91s/it] {'loss': 0.1624, 'grad_norm': 0.16971106827259064, 'learning_rate': 0.00010196975050515315, 'epoch': 3.69} + 74%|███████▎ | 553/750 [57:38<19:23, 5.91s/it] 74%|███████▍ | 554/750 [57:44<19:20, 5.92s/it] {'loss': 0.147, 'grad_norm': 0.15760289132595062, 'learning_rate': 0.00010187767847689286, 'epoch': 3.69} + 74%|███████▍ | 554/750 [57:44<19:20, 5.92s/it] 74%|███████▍ | 555/750 [57:50<19:11, 5.91s/it] {'loss': 0.1293, 'grad_norm': 0.14435075223445892, 'learning_rate': 0.00010178585540417863, 'epoch': 3.7} + 74%|███████▍ | 555/750 [57:50<19:11, 5.91s/it] 74%|███████▍ | 556/750 [57:56<19:06, 5.91s/it] {'loss': 0.1524, 'grad_norm': 0.1684872806072235, 'learning_rate': 0.00010169428016710201, 'epoch': 3.71} + 74%|███████▍ | 556/750 [57:56<19:06, 5.91s/it] 74%|███████▍ | 557/750 [58:02<18:59, 5.90s/it] {'loss': 0.1433, 'grad_norm': 0.14643168449401855, 'learning_rate': 0.00010160295165279474, 'epoch': 3.71} + 74%|███████▍ | 557/750 [58:02<18:59, 5.90s/it] 74%|███████▍ | 558/750 [58:08<18:50, 5.89s/it] {'loss': 0.1374, 'grad_norm': 0.16298243403434753, 'learning_rate': 0.00010151186875537208, 'epoch': 3.72} + 74%|███████▍ | 558/750 [58:08<18:50, 5.89s/it] 75%|███████▍ | 559/750 [58:14<18:44, 5.89s/it] {'loss': 0.1472, 'grad_norm': 0.1614677906036377, 'learning_rate': 0.00010142103037587655, 'epoch': 3.73} + 75%|███████▍ | 559/750 [58:14<18:44, 5.89s/it] 75%|███████▍ | 560/750 [58:20<18:50, 5.95s/it] {'loss': 0.1409, 'grad_norm': 0.15404939651489258, 'learning_rate': 0.00010133043542222218, 'epoch': 3.73} + 75%|███████▍ | 560/750 [58:20<18:50, 5.95s/it] 75%|███████▍ | 561/750 [58:26<18:45, 5.95s/it] {'loss': 0.1418, 'grad_norm': 0.1949167400598526, 'learning_rate': 0.00010124008280913954, 'epoch': 3.74} + 75%|███████▍ | 561/750 [58:26<18:45, 5.95s/it] 75%|███████▍ | 562/750 [58:32<18:42, 5.97s/it] {'loss': 0.1271, 'grad_norm': 0.15182584524154663, 'learning_rate': 0.00010114997145812079, 'epoch': 3.75} + 75%|███████▍ | 562/750 [58:32<18:42, 5.97s/it] 75%|███████▌ | 563/750 [58:38<18:32, 5.95s/it] {'loss': 0.142, 'grad_norm': 0.17694354057312012, 'learning_rate': 0.000101060100297366, 'epoch': 3.75} + 75%|███████▌ | 563/750 [58:38<18:32, 5.95s/it] 75%|███████▌ | 564/750 [58:44<18:24, 5.94s/it] {'loss': 0.1376, 'grad_norm': 0.16230152547359467, 'learning_rate': 0.00010097046826172943, 'epoch': 3.76} + 75%|███████▌ | 564/750 [58:44<18:24, 5.94s/it] 75%|███████▌ | 565/750 [58:50<18:21, 5.96s/it] {'loss': 0.1398, 'grad_norm': 0.15251514315605164, 'learning_rate': 0.0001008810742926664, 'epoch': 3.77} + 75%|███████▌ | 565/750 [58:50<18:21, 5.96s/it] 75%|███████▌ | 566/750 [58:56<18:12, 5.94s/it] {'loss': 0.1561, 'grad_norm': 0.16820892691612244, 'learning_rate': 0.00010079191733818102, 'epoch': 3.77} + 75%|███████▌ | 566/750 [58:56<18:12, 5.94s/it] 76%|███████▌ | 567/750 [59:02<18:07, 5.94s/it] {'loss': 0.1633, 'grad_norm': 0.1608879268169403, 'learning_rate': 0.00010070299635277398, 'epoch': 3.78} + 76%|███████▌ | 567/750 [59:02<18:07, 5.94s/it] 76%|███████▌ | 568/750 [59:08<18:04, 5.96s/it] {'loss': 0.1519, 'grad_norm': 0.16159476339817047, 'learning_rate': 0.00010061431029739118, 'epoch': 3.79} + 76%|███████▌ | 568/750 [59:08<18:04, 5.96s/it] 76%|███████▌ | 569/750 [59:14<18:02, 5.98s/it] {'loss': 0.1392, 'grad_norm': 0.150252103805542, 'learning_rate': 0.00010052585813937274, 'epoch': 3.79} + 76%|███████▌ | 569/750 [59:14<18:02, 5.98s/it] 76%|███████▌ | 570/750 [59:19<17:52, 5.96s/it] {'loss': 0.1627, 'grad_norm': 0.1768335998058319, 'learning_rate': 0.0001004376388524024, 'epoch': 3.8} + 76%|███████▌ | 570/750 [59:19<17:52, 5.96s/it] 76%|███████▌ | 571/750 [59:25<17:42, 5.94s/it] {'loss': 0.1469, 'grad_norm': 0.15728168189525604, 'learning_rate': 0.00010034965141645762, 'epoch': 3.81} + 76%|███████▌ | 571/750 [59:25<17:42, 5.94s/it] 76%|███████▋ | 572/750 [59:31<17:30, 5.90s/it] {'loss': 0.1472, 'grad_norm': 0.156614288687706, 'learning_rate': 0.00010026189481775992, 'epoch': 3.81} + 76%|███████▋ | 572/750 [59:31<17:30, 5.90s/it] 76%|███████▋ | 573/750 [59:37<17:26, 5.91s/it] {'loss': 0.1467, 'grad_norm': 0.16038380563259125, 'learning_rate': 0.00010017436804872594, 'epoch': 3.82} + 76%|███████▋ | 573/750 [59:37<17:26, 5.91s/it] 77%|███████▋ | 574/750 [59:43<17:16, 5.89s/it] {'loss': 0.1417, 'grad_norm': 0.16885799169540405, 'learning_rate': 0.00010008707010791883, 'epoch': 3.83} + 77%|███████▋ | 574/750 [59:43<17:16, 5.89s/it] 77%|███████▋ | 575/750 [59:49<17:12, 5.90s/it] {'loss': 0.1533, 'grad_norm': 0.15518838167190552, 'learning_rate': 0.0001, 'epoch': 3.83} + 77%|███████▋ | 575/750 [59:49<17:12, 5.90s/it] 77%|███████▋ | 576/750 [59:55<17:03, 5.88s/it] {'loss': 0.1409, 'grad_norm': 0.16575117409229279, 'learning_rate': 9.991315673568166e-05, 'epoch': 3.84} + 77%|███████▋ | 576/750 [59:55<17:03, 5.88s/it] 77%|███████▋ | 577/750 [1:00:01<16:59, 5.89s/it] {'loss': 0.1452, 'grad_norm': 0.16730660200119019, 'learning_rate': 9.982653933167944e-05, 'epoch': 3.85} + 77%|███████▋ | 577/750 [1:00:01<16:59, 5.89s/it] 77%|███████▋ | 578/750 [1:00:07<16:52, 5.89s/it] {'loss': 0.1584, 'grad_norm': 0.1649489849805832, 'learning_rate': 9.97401468106657e-05, 'epoch': 3.85} + 77%|███████▋ | 578/750 [1:00:07<16:52, 5.89s/it] 77%|███████▋ | 579/750 [1:00:13<16:59, 5.96s/it] {'loss': 0.1497, 'grad_norm': 0.16405245661735535, 'learning_rate': 9.965397820122326e-05, 'epoch': 3.86} + 77%|███████▋ | 579/750 [1:00:13<16:59, 5.96s/it] 77%|███████▋ | 580/750 [1:00:19<16:54, 5.97s/it] {'loss': 0.1303, 'grad_norm': 0.15870530903339386, 'learning_rate': 9.956803253779939e-05, 'epoch': 3.87} + 77%|███████▋ | 580/750 [1:00:19<16:54, 5.97s/it] 77%|███████▋ | 581/750 [1:00:25<16:45, 5.95s/it] {'loss': 0.1469, 'grad_norm': 0.16256536543369293, 'learning_rate': 9.94823088606605e-05, 'epoch': 3.87} + 77%|███████▋ | 581/750 [1:00:25<16:45, 5.95s/it] 78%|███████▊ | 582/750 [1:00:30<16:36, 5.93s/it] {'loss': 0.1441, 'grad_norm': 0.1538584679365158, 'learning_rate': 9.939680621584706e-05, 'epoch': 3.88} + 78%|███████▊ | 582/750 [1:00:30<16:36, 5.93s/it] 78%|███████▊ | 583/750 [1:00:36<16:31, 5.94s/it] {'loss': 0.163, 'grad_norm': 0.1761755496263504, 'learning_rate': 9.931152365512906e-05, 'epoch': 3.89} + 78%|███████▊ | 583/750 [1:00:36<16:31, 5.94s/it] 78%|███████▊ | 584/750 [1:00:42<16:26, 5.95s/it] {'loss': 0.1463, 'grad_norm': 0.1586264669895172, 'learning_rate': 9.922646023596178e-05, 'epoch': 3.89} + 78%|███████▊ | 584/750 [1:00:42<16:26, 5.95s/it] 78%|███████▊ | 585/750 [1:00:48<16:18, 5.93s/it] {'loss': 0.1343, 'grad_norm': 0.16324672102928162, 'learning_rate': 9.914161502144209e-05, 'epoch': 3.9} + 78%|███████▊ | 585/750 [1:00:48<16:18, 5.93s/it] 78%|███████▊ | 586/750 [1:00:54<16:11, 5.93s/it] {'loss': 0.1528, 'grad_norm': 0.17567071318626404, 'learning_rate': 9.905698708026503e-05, 'epoch': 3.91} + 78%|███████▊ | 586/750 [1:00:54<16:11, 5.93s/it] 78%|███████▊ | 587/750 [1:01:00<16:12, 5.97s/it] {'loss': 0.1512, 'grad_norm': 0.15445156395435333, 'learning_rate': 9.897257548668106e-05, 'epoch': 3.91} + 78%|███████▊ | 587/750 [1:01:00<16:12, 5.97s/it] 78%|███████▊ | 588/750 [1:01:06<16:02, 5.94s/it] {'loss': 0.1334, 'grad_norm': 0.15348082780838013, 'learning_rate': 9.888837932045328e-05, 'epoch': 3.92} + 78%|███████▊ | 588/750 [1:01:06<16:02, 5.94s/it] 79%|███████▊ | 589/750 [1:01:12<15:56, 5.94s/it] {'loss': 0.1751, 'grad_norm': 0.16937798261642456, 'learning_rate': 9.880439766681545e-05, 'epoch': 3.93} + 79%|███████▊ | 589/750 [1:01:12<15:56, 5.94s/it] 79%|███████▊ | 590/750 [1:01:18<15:52, 5.95s/it] {'loss': 0.1494, 'grad_norm': 0.16821295022964478, 'learning_rate': 9.872062961643026e-05, 'epoch': 3.93} + 79%|███████▊ | 590/750 [1:01:18<15:52, 5.95s/it] 79%|███████▉ | 591/750 [1:01:24<15:41, 5.92s/it] {'loss': 0.1591, 'grad_norm': 0.17528890073299408, 'learning_rate': 9.863707426534786e-05, 'epoch': 3.94} + 79%|███████▉ | 591/750 [1:01:24<15:41, 5.92s/it] 79%|███████▉ | 592/750 [1:01:30<15:32, 5.90s/it] {'loss': 0.1575, 'grad_norm': 0.17449085414409637, 'learning_rate': 9.8553730714965e-05, 'epoch': 3.95} + 79%|███████▉ | 592/750 [1:01:30<15:32, 5.90s/it] 79%|███████▉ | 593/750 [1:01:36<15:27, 5.91s/it] {'loss': 0.1634, 'grad_norm': 0.1672595590353012, 'learning_rate': 9.847059807198442e-05, 'epoch': 3.95} + 79%|███████▉ | 593/750 [1:01:36<15:27, 5.91s/it] 79%|███████▉ | 594/750 [1:01:42<15:25, 5.94s/it] {'loss': 0.1698, 'grad_norm': 0.16503292322158813, 'learning_rate': 9.838767544837452e-05, 'epoch': 3.96} + 79%|███████▉ | 594/750 [1:01:42<15:25, 5.94s/it] 79%|███████▉ | 595/750 [1:01:47<15:14, 5.90s/it] {'loss': 0.1483, 'grad_norm': 0.17282924056053162, 'learning_rate': 9.830496196132975e-05, 'epoch': 3.97} + 79%|███████▉ | 595/750 [1:01:47<15:14, 5.90s/it] 79%|███████▉ | 596/750 [1:01:53<15:13, 5.93s/it] {'loss': 0.1495, 'grad_norm': 0.15530723333358765, 'learning_rate': 9.822245673323086e-05, 'epoch': 3.97} + 79%|███████▉ | 596/750 [1:01:53<15:13, 5.93s/it] 80%|███████▉ | 597/750 [1:02:00<15:15, 5.99s/it] {'loss': 0.1408, 'grad_norm': 0.1515883356332779, 'learning_rate': 9.814015889160605e-05, 'epoch': 3.98} + 80%|███████▉ | 597/750 [1:02:00<15:15, 5.99s/it] 80%|███████▉ | 598/750 [1:02:05<15:05, 5.96s/it] {'loss': 0.1486, 'grad_norm': 0.1659146249294281, 'learning_rate': 9.805806756909202e-05, 'epoch': 3.99} + 80%|███████▉ | 598/750 [1:02:05<15:05, 5.96s/it] 80%|███████▉ | 599/750 [1:02:11<15:00, 5.96s/it] {'loss': 0.1563, 'grad_norm': 0.1595316231250763, 'learning_rate': 9.797618190339569e-05, 'epoch': 3.99} + 80%|███████▉ | 599/750 [1:02:11<15:00, 5.96s/it] 80%|████████ | 600/750 [1:02:20<16:31, 6.61s/it] {'loss': 0.1357, 'grad_norm': 0.1535337269306183, 'learning_rate': 9.789450103725609e-05, 'epoch': 4.0} + 80%|████████ | 600/750 [1:02:20<16:31, 6.61s/it][INFO|trainer.py:3831] 2025-06-27 00:51:07,062 >> +***** Running Evaluation ***** +[INFO|trainer.py:3833] 2025-06-27 00:51:07,062 >> Num examples = 1000 +[INFO|trainer.py:3836] 2025-06-27 00:51:07,062 >> Batch size = 25 + + 0%| | 0/10 [00:00> Saving model checkpoint to ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-600 +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9dce-3f19ec562c8ad0eb1015a041;7fd38569-953b-4fc4-9e3c-f64d3d33a99a) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +[INFO|tokenization_utils_base.py:2684] 2025-06-27 00:51:51,144 >> tokenizer config file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-600/tokenizer_config.json +[INFO|tokenization_utils_base.py:2693] 2025-06-27 00:51:51,144 >> Special tokens file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-600/special_tokens_map.json +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user. + warnings.warn( # warn only once +[2025-06-27 00:51:53,843] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step600 is begin to save! +[2025-06-27 00:51:53,866] [INFO] [logging.py:107:log_dist] [Rank 0] Saving model checkpoint: ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-600/global_step600/mp_rank_00_model_states.pt +[INFO|trainer.py:3607] 2025-06-27 00:51:53,992 >> Deleting older checkpoint [outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-200] due to args.save_total_limit +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9dd2-3b36755858a1faf512f6af73;2910ee7c-1f93-404a-a256-b34106f8c3dc) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9dd2-470bb0bd7346b0fa114344dc;29d37579-b0aa-46a9-b0ec-450107b9a45c) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9dd2-07f6408654e3f9292492f746;a4377ecd-2002-4607-8a3a-e413c0ff315f) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685d9dd2-33813d07324bd09b6b4b232b;03bee699-db56-41b8-9aa2-f89c0b83e37f) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/save_and_load.py:236: UserWarning: Could not find a config file in meta-llama/Llama-3.1-8B-Instruct - will assume that the vocabulary was not modified. + warnings.warn( + 80%|████████ | 601/750 [1:03:21<57:30, 23.16s/it] {'loss': 0.1297, 'grad_norm': 0.16146820783615112, 'learning_rate': 9.781302411840674e-05, 'epoch': 4.01} + 80%|████████ | 601/750 [1:03:21<57:30, 23.16s/it] 80%|████████ | 602/750 [1:03:27<44:21, 17.99s/it] {'loss': 0.1294, 'grad_norm': 0.15809302031993866, 'learning_rate': 9.773175029953825e-05, 'epoch': 4.01} + 80%|████████ | 602/750 [1:03:27<44:21, 17.99s/it] 80%|████████ | 603/750 [1:03:33<35:07, 14.33s/it] {'loss': 0.109, 'grad_norm': 0.15633410215377808, 'learning_rate': 9.76506787382613e-05, 'epoch': 4.02} + 80%|████████ | 603/750 [1:03:33<35:07, 14.33s/it] 81%|████████ | 604/750 [1:03:39<28:45, 11.82s/it] {'loss': 0.1251, 'grad_norm': 0.1713617891073227, 'learning_rate': 9.756980859707e-05, 'epoch': 4.03} + 81%|████████ | 604/750 [1:03:39<28:45, 11.82s/it] 81%|████████ | 605/750 [1:03:45<24:19, 10.06s/it] {'loss': 0.1219, 'grad_norm': 0.19035956263542175, 'learning_rate': 9.748913904330553e-05, 'epoch': 4.03} + 81%|████████ | 605/750 [1:03:45<24:19, 10.06s/it] 81%|████████ | 606/750 [1:03:51<21:13, 8.84s/it] {'loss': 0.1209, 'grad_norm': 0.20892535150051117, 'learning_rate': 9.740866924912017e-05, 'epoch': 4.04} + 81%|████████ | 606/750 [1:03:51<21:13, 8.84s/it] 81%|████████ | 607/750 [1:03:57<19:04, 8.00s/it] {'loss': 0.1172, 'grad_norm': 0.20441478490829468, 'learning_rate': 9.732839839144154e-05, 'epoch': 4.05} + 81%|████████ | 607/750 [1:03:57<19:04, 8.00s/it] 81%|████████ | 608/750 [1:04:03<17:29, 7.39s/it] {'loss': 0.1122, 'grad_norm': 0.20807655155658722, 'learning_rate': 9.724832565193738e-05, 'epoch': 4.05} + 81%|████████ | 608/750 [1:04:03<17:29, 7.39s/it] 81%|████████ | 609/750 [1:04:09<16:22, 6.97s/it] {'loss': 0.1208, 'grad_norm': 0.18891528248786926, 'learning_rate': 9.716845021698033e-05, 'epoch': 4.06} + 81%|████████ | 609/750 [1:04:09<16:22, 6.97s/it] 81%|████████▏ | 610/750 [1:04:15<15:32, 6.66s/it] {'loss': 0.1073, 'grad_norm': 0.17771846055984497, 'learning_rate': 9.708877127761337e-05, 'epoch': 4.07} + 81%|████████▏ | 610/750 [1:04:15<15:32, 6.66s/it] 81%|████████▏ | 611/750 [1:04:21<14:57, 6.46s/it] {'loss': 0.1144, 'grad_norm': 0.16970933973789215, 'learning_rate': 9.700928802951527e-05, 'epoch': 4.07} + 81%|████████▏ | 611/750 [1:04:21<14:57, 6.46s/it] 82%|████████▏ | 612/750 [1:04:27<14:31, 6.31s/it] {'loss': 0.1232, 'grad_norm': 0.18293039500713348, 'learning_rate': 9.69299996729666e-05, 'epoch': 4.08} + 82%|████████▏ | 612/750 [1:04:27<14:31, 6.31s/it] 82%|████████▏ | 613/750 [1:04:33<14:10, 6.21s/it] {'loss': 0.1247, 'grad_norm': 0.172981396317482, 'learning_rate': 9.685090541281594e-05, 'epoch': 4.09} + 82%|████████▏ | 613/750 [1:04:33<14:10, 6.21s/it] 82%|████████▏ | 614/750 [1:04:39<13:56, 6.15s/it] {'loss': 0.1178, 'grad_norm': 0.16729474067687988, 'learning_rate': 9.677200445844624e-05, 'epoch': 4.09} + 82%|████████▏ | 614/750 [1:04:39<13:56, 6.15s/it] 82%|████████▏ | 615/750 [1:04:45<13:43, 6.10s/it] {'loss': 0.1212, 'grad_norm': 0.16877005994319916, 'learning_rate': 9.669329602374179e-05, 'epoch': 4.1} + 82%|████████▏ | 615/750 [1:04:45<13:43, 6.10s/it] 82%|████████▏ | 616/750 [1:04:51<13:42, 6.14s/it] {'loss': 0.1186, 'grad_norm': 0.1810474544763565, 'learning_rate': 9.661477932705526e-05, 'epoch': 4.11} + 82%|████████▏ | 616/750 [1:04:51<13:42, 6.14s/it] 82%|████████▏ | 617/750 [1:04:57<13:31, 6.10s/it] {'loss': 0.138, 'grad_norm': 0.1953781545162201, 'learning_rate': 9.65364535911751e-05, 'epoch': 4.11} + 82%|████████▏ | 617/750 [1:04:57<13:31, 6.10s/it] 82%|████████▏ | 618/750 [1:05:03<13:27, 6.11s/it] {'loss': 0.107, 'grad_norm': 0.1810551881790161, 'learning_rate': 9.645831804329319e-05, 'epoch': 4.12} + 82%|████████▏ | 618/750 [1:05:03<13:27, 6.11s/it] 83%|████████▎ | 619/750 [1:05:09<13:12, 6.05s/it] {'loss': 0.1258, 'grad_norm': 0.21333763003349304, 'learning_rate': 9.63803719149729e-05, 'epoch': 4.13} + 83%|████████▎ | 619/750 [1:05:09<13:12, 6.05s/it] 83%|████████▎ | 620/750 [1:05:15<12:49, 5.92s/it] {'loss': 0.1135, 'grad_norm': 0.1991916447877884, 'learning_rate': 9.630261444211729e-05, 'epoch': 4.13} + 83%|████████▎ | 620/750 [1:05:15<12:49, 5.92s/it] 83%|████████▎ | 621/750 [1:05:21<12:43, 5.92s/it] {'loss': 0.1283, 'grad_norm': 0.19787324965000153, 'learning_rate': 9.622504486493763e-05, 'epoch': 4.14} + 83%|████████▎ | 621/750 [1:05:21<12:43, 5.92s/it] 83%|████████▎ | 622/750 [1:05:27<12:40, 5.95s/it] {'loss': 0.1085, 'grad_norm': 0.17439593374729156, 'learning_rate': 9.614766242792228e-05, 'epoch': 4.15} + 83%|████████▎ | 622/750 [1:05:27<12:40, 5.95s/it] 83%|████████▎ | 623/750 [1:05:33<12:35, 5.95s/it] {'loss': 0.1133, 'grad_norm': 0.1835867315530777, 'learning_rate': 9.607046637980575e-05, 'epoch': 4.15} + 83%|████████▎ | 623/750 [1:05:33<12:35, 5.95s/it] 83%|████████▎ | 624/750 [1:05:39<12:28, 5.94s/it] {'loss': 0.1113, 'grad_norm': 0.1911616176366806, 'learning_rate': 9.599345597353806e-05, 'epoch': 4.16} + 83%|████████▎ | 624/750 [1:05:39<12:28, 5.94s/it] 83%|████████▎ | 625/750 [1:05:45<12:23, 5.95s/it] {'loss': 0.1128, 'grad_norm': 0.19186195731163025, 'learning_rate': 9.59166304662544e-05, 'epoch': 4.17} + 83%|████████▎ | 625/750 [1:05:45<12:23, 5.95s/it] 83%|████████▎ | 626/750 [1:05:50<12:18, 5.95s/it] {'loss': 0.1317, 'grad_norm': 0.20234230160713196, 'learning_rate': 9.5839989119245e-05, 'epoch': 4.17} + 83%|████████▎ | 626/750 [1:05:50<12:18, 5.95s/it] 84%|████████▎ | 627/750 [1:05:56<12:09, 5.93s/it] {'loss': 0.1157, 'grad_norm': 0.1942509412765503, 'learning_rate': 9.57635311979253e-05, 'epoch': 4.18} + 84%|████████▎ | 627/750 [1:05:56<12:09, 5.93s/it] 84%|████████▎ | 628/750 [1:06:02<12:02, 5.92s/it] {'loss': 0.1217, 'grad_norm': 0.1894557774066925, 'learning_rate': 9.568725597180648e-05, 'epoch': 4.19} + 84%|████████▎ | 628/750 [1:06:02<12:02, 5.92s/it] 84%|████████▍ | 629/750 [1:06:08<12:00, 5.96s/it] {'loss': 0.0997, 'grad_norm': 0.16736949980258942, 'learning_rate': 9.561116271446594e-05, 'epoch': 4.19} + 84%|████████▍ | 629/750 [1:06:08<12:00, 5.96s/it] 84%|████████▍ | 630/750 [1:06:14<11:56, 5.97s/it] {'loss': 0.1085, 'grad_norm': 0.1881280392408371, 'learning_rate': 9.553525070351848e-05, 'epoch': 4.2} + 84%|████████▍ | 630/750 [1:06:14<11:56, 5.97s/it] 84%|████████▍ | 631/750 [1:06:20<11:51, 5.98s/it] {'loss': 0.1052, 'grad_norm': 0.19926777482032776, 'learning_rate': 9.54595192205873e-05, 'epoch': 4.21} + 84%|████████▍ | 631/750 [1:06:20<11:51, 5.98s/it] 84%|████████▍ | 632/750 [1:06:26<11:48, 6.00s/it] {'loss': 0.1249, 'grad_norm': 0.19126403331756592, 'learning_rate': 9.538396755127553e-05, 'epoch': 4.21} + 84%|████████▍ | 632/750 [1:06:26<11:48, 6.00s/it] 84%|████████▍ | 633/750 [1:06:32<11:39, 5.98s/it] {'loss': 0.1417, 'grad_norm': 0.2144770473241806, 'learning_rate': 9.530859498513802e-05, 'epoch': 4.22} + 84%|████████▍ | 633/750 [1:06:32<11:39, 5.98s/it] 85%|████████▍ | 634/750 [1:06:38<11:39, 6.03s/it] {'loss': 0.1104, 'grad_norm': 0.20596463978290558, 'learning_rate': 9.523340081565305e-05, 'epoch': 4.23} + 85%|████████▍ | 634/750 [1:06:38<11:39, 6.03s/it] 85%|████████▍ | 635/750 [1:06:44<11:32, 6.02s/it] {'loss': 0.1268, 'grad_norm': 0.19861352443695068, 'learning_rate': 9.515838434019473e-05, 'epoch': 4.23} + 85%|████████▍ | 635/750 [1:06:44<11:32, 6.02s/it] 85%|████████▍ | 636/750 [1:06:50<11:22, 5.99s/it] {'loss': 0.1056, 'grad_norm': 0.1768050193786621, 'learning_rate': 9.508354486000538e-05, 'epoch': 4.24} + 85%|████████▍ | 636/750 [1:06:50<11:22, 5.99s/it] 85%|████████▍ | 637/750 [1:06:56<11:21, 6.03s/it] {'loss': 0.1095, 'grad_norm': 0.17600397765636444, 'learning_rate': 9.500888168016802e-05, 'epoch': 4.25} + 85%|████████▍ | 637/750 [1:06:56<11:21, 6.03s/it] 85%|████████▌ | 638/750 [1:07:02<11:11, 6.00s/it] {'loss': 0.1223, 'grad_norm': 0.1911844164133072, 'learning_rate': 9.493439410957959e-05, 'epoch': 4.25} + 85%|████████▌ | 638/750 [1:07:02<11:11, 6.00s/it] 85%|████████▌ | 639/750 [1:07:08<11:01, 5.96s/it] {'loss': 0.1287, 'grad_norm': 0.20859798789024353, 'learning_rate': 9.48600814609237e-05, 'epoch': 4.26} + 85%|████████▌ | 639/750 [1:07:08<11:01, 5.96s/it] 85%|████████▌ | 640/750 [1:07:14<10:55, 5.96s/it] {'loss': 0.1273, 'grad_norm': 0.19483715295791626, 'learning_rate': 9.478594305064439e-05, 'epoch': 4.27} + 85%|████████▌ | 640/750 [1:07:14<10:55, 5.96s/it] 85%|████████▌ | 641/750 [1:07:20<10:50, 5.97s/it] {'loss': 0.1261, 'grad_norm': 0.19489359855651855, 'learning_rate': 9.471197819891936e-05, 'epoch': 4.27} + 85%|████████▌ | 641/750 [1:07:20<10:50, 5.97s/it] 86%|████████▌ | 642/750 [1:07:26<10:39, 5.92s/it] {'loss': 0.1275, 'grad_norm': 0.21312512457370758, 'learning_rate': 9.463818622963402e-05, 'epoch': 4.28} + 86%|████████▌ | 642/750 [1:07:26<10:39, 5.92s/it] 86%|████████▌ | 643/750 [1:07:32<10:34, 5.93s/it] {'loss': 0.1192, 'grad_norm': 0.21053853631019592, 'learning_rate': 9.45645664703555e-05, 'epoch': 4.29} + 86%|████████▌ | 643/750 [1:07:32<10:34, 5.93s/it] 86%|████████▌ | 644/750 [1:07:38<10:29, 5.94s/it] {'loss': 0.1274, 'grad_norm': 0.19626310467720032, 'learning_rate': 9.44911182523068e-05, 'epoch': 4.29} + 86%|████████▌ | 644/750 [1:07:38<10:29, 5.94s/it] 86%|████████▌ | 645/750 [1:07:44<10:24, 5.94s/it] {'loss': 0.1096, 'grad_norm': 0.19516772031784058, 'learning_rate': 9.441784091034141e-05, 'epoch': 4.3} + 86%|████████▌ | 645/750 [1:07:44<10:24, 5.94s/it] 86%|████████▌ | 646/750 [1:07:50<10:20, 5.97s/it] {'loss': 0.1203, 'grad_norm': 0.1724524199962616, 'learning_rate': 9.434473378291795e-05, 'epoch': 4.31} + 86%|████████▌ | 646/750 [1:07:50<10:20, 5.97s/it] 86%|████████▋ | 647/750 [1:07:56<10:12, 5.95s/it] {'loss': 0.118, 'grad_norm': 0.18564382195472717, 'learning_rate': 9.427179621207505e-05, 'epoch': 4.31} + 86%|████████▋ | 647/750 [1:07:56<10:12, 5.95s/it] 86%|████████▋ | 648/750 [1:08:02<10:08, 5.97s/it] {'loss': 0.1329, 'grad_norm': 0.198990136384964, 'learning_rate': 9.419902754340649e-05, 'epoch': 4.32} + 86%|████████▋ | 648/750 [1:08:02<10:08, 5.97s/it] 87%|████████▋ | 649/750 [1:08:08<10:03, 5.98s/it] {'loss': 0.1229, 'grad_norm': 0.19019605219364166, 'learning_rate': 9.41264271260367e-05, 'epoch': 4.33} + 87%|████████▋ | 649/750 [1:08:08<10:03, 5.98s/it] 87%|████████▋ | 650/750 [1:08:14<09:56, 5.97s/it] {'loss': 0.1068, 'grad_norm': 0.1980881541967392, 'learning_rate': 9.405399431259603e-05, 'epoch': 4.33} + 87%|████████▋ | 650/750 [1:08:14<09:56, 5.97s/it] 87%|████████▋ | 651/750 [1:08:20<09:48, 5.95s/it] {'loss': 0.125, 'grad_norm': 0.19980546832084656, 'learning_rate': 9.398172845919671e-05, 'epoch': 4.34} + 87%|████████▋ | 651/750 [1:08:20<09:48, 5.95s/it] 87%|████████▋ | 652/750 [1:08:26<09:42, 5.94s/it] {'loss': 0.1274, 'grad_norm': 0.20640608668327332, 'learning_rate': 9.390962892540881e-05, 'epoch': 4.35} + 87%|████████▋ | 652/750 [1:08:26<09:42, 5.94s/it] 87%|████████▋ | 653/750 [1:08:31<09:33, 5.91s/it] {'loss': 0.1215, 'grad_norm': 0.20882228016853333, 'learning_rate': 9.383769507423618e-05, 'epoch': 4.35} + 87%|████████▋ | 653/750 [1:08:31<09:33, 5.91s/it] 87%|████████▋ | 654/750 [1:08:37<09:28, 5.92s/it] {'loss': 0.1367, 'grad_norm': 0.2102683037519455, 'learning_rate': 9.376592627209315e-05, 'epoch': 4.36} + 87%|████████▋ | 654/750 [1:08:37<09:28, 5.92s/it] 87%|███��████▋ | 655/750 [1:08:43<09:26, 5.96s/it] {'loss': 0.107, 'grad_norm': 0.18556328117847443, 'learning_rate': 9.369432188878082e-05, 'epoch': 4.37} + 87%|████████▋ | 655/750 [1:08:43<09:26, 5.96s/it] 87%|████████▋ | 656/750 [1:08:49<09:16, 5.92s/it] {'loss': 0.1162, 'grad_norm': 0.19150245189666748, 'learning_rate': 9.362288129746395e-05, 'epoch': 4.37} + 87%|████████▋ | 656/750 [1:08:49<09:16, 5.92s/it] 88%|████████▊ | 657/750 [1:08:55<09:10, 5.92s/it] {'loss': 0.1211, 'grad_norm': 0.19354361295700073, 'learning_rate': 9.355160387464784e-05, 'epoch': 4.38} + 88%|████████▊ | 657/750 [1:08:55<09:10, 5.92s/it] 88%|████████▊ | 658/750 [1:09:01<09:05, 5.93s/it] {'loss': 0.1227, 'grad_norm': 0.18872122466564178, 'learning_rate': 9.348048900015556e-05, 'epoch': 4.39} + 88%|████████▊ | 658/750 [1:09:01<09:05, 5.93s/it] 88%|████████▊ | 659/750 [1:09:07<08:59, 5.93s/it] {'loss': 0.1156, 'grad_norm': 0.1734706163406372, 'learning_rate': 9.340953605710524e-05, 'epoch': 4.39} + 88%|████████▊ | 659/750 [1:09:07<08:59, 5.93s/it] 88%|████████▊ | 660/750 [1:09:13<08:52, 5.91s/it] {'loss': 0.1321, 'grad_norm': 0.19426502287387848, 'learning_rate': 9.333874443188752e-05, 'epoch': 4.4} + 88%|████████▊ | 660/750 [1:09:13<08:52, 5.91s/it] 88%|████████▊ | 661/750 [1:09:19<08:48, 5.94s/it] {'loss': 0.1268, 'grad_norm': 0.18161000311374664, 'learning_rate': 9.326811351414343e-05, 'epoch': 4.41} + 88%|████████▊ | 661/750 [1:09:19<08:48, 5.94s/it] 88%|████████▊ | 662/750 [1:09:25<08:44, 5.96s/it] {'loss': 0.1154, 'grad_norm': 0.18544411659240723, 'learning_rate': 9.319764269674211e-05, 'epoch': 4.41} + 88%|████████▊ | 662/750 [1:09:25<08:44, 5.96s/it] 88%|████████▊ | 663/750 [1:09:31<08:34, 5.92s/it] {'loss': 0.1313, 'grad_norm': 0.1945258527994156, 'learning_rate': 9.312733137575907e-05, 'epoch': 4.42} + 88%|████████▊ | 663/750 [1:09:31<08:34, 5.92s/it] 89%|████████▊ | 664/750 [1:09:37<08:30, 5.93s/it] {'loss': 0.1286, 'grad_norm': 0.18810269236564636, 'learning_rate': 9.305717895045427e-05, 'epoch': 4.43} + 89%|████████▊ | 664/750 [1:09:37<08:30, 5.93s/it] 89%|████████▊ | 665/750 [1:09:43<08:25, 5.94s/it] {'loss': 0.1259, 'grad_norm': 0.20028351247310638, 'learning_rate': 9.298718482325068e-05, 'epoch': 4.43} + 89%|████████▊ | 665/750 [1:09:43<08:25, 5.94s/it] 89%|████████▉ | 666/750 [1:09:49<08:18, 5.94s/it] {'loss': 0.1435, 'grad_norm': 0.2324310839176178, 'learning_rate': 9.291734839971293e-05, 'epoch': 4.44} + 89%|████████▉ | 666/750 [1:09:49<08:18, 5.94s/it] 89%|████████▉ | 667/750 [1:09:55<08:13, 5.94s/it] {'loss': 0.1313, 'grad_norm': 0.20299063622951508, 'learning_rate': 9.284766908852593e-05, 'epoch': 4.45} + 89%|████████▉ | 667/750 [1:09:55<08:13, 5.94s/it] 89%|████████▉ | 668/750 [1:10:00<08:05, 5.92s/it] {'loss': 0.1149, 'grad_norm': 0.20010414719581604, 'learning_rate': 9.277814630147407e-05, 'epoch': 4.45} + 89%|████████▉ | 668/750 [1:10:00<08:05, 5.92s/it] 89%|████████▉ | 669/750 [1:10:06<07:58, 5.91s/it] {'loss': 0.127, 'grad_norm': 0.20554889738559723, 'learning_rate': 9.270877945342015e-05, 'epoch': 4.46} + 89%|████████▉ | 669/750 [1:10:06<07:58, 5.91s/it] 89%|████████▉ | 670/750 [1:10:12<07:53, 5.91s/it] {'loss': 0.1151, 'grad_norm': 0.19019319117069244, 'learning_rate': 9.263956796228493e-05, 'epoch': 4.47} + 89%|████████▉ | 670/750 [1:10:12<07:53, 5.91s/it] 89%|████████▉ | 671/750 [1:10:18<07:45, 5.90s/it] {'loss': 0.1133, 'grad_norm': 0.1894761323928833, 'learning_rate': 9.257051124902634e-05, 'epoch': 4.47} + 89%|████████▉ | 671/750 [1:10:18<07:45, 5.90s/it] 90%|████████▉ | 672/750 [1:10:24<07:41, 5.92s/it] {'loss': 0.1227, 'grad_norm': 0.21622046828269958, 'learning_rate': 9.250160873761937e-05, 'epoch': 4.48} + 90%|████████▉ | 672/750 [1:10:24<07:41, 5.92s/it] 90%|████████▉ | 673/750 [1:10:30<07:37, 5.94s/it] {'loss': 0.1096, 'grad_norm': 0.18021948635578156, 'learning_rate': 9.243285985503578e-05, 'epoch': 4.49} + 90%|████████▉ | 673/750 [1:10:30<07:37, 5.94s/it] 90%|████████▉ | 674/750 [1:10:36<07:33, 5.97s/it] {'loss': 0.1162, 'grad_norm': 0.18876951932907104, 'learning_rate': 9.236426403122409e-05, 'epoch': 4.49} + 90%|████████▉ | 674/750 [1:10:36<07:33, 5.97s/it] 90%|█████████ | 675/750 [1:10:42<07:25, 5.94s/it] {'loss': 0.1111, 'grad_norm': 0.19144152104854584, 'learning_rate': 9.229582069908974e-05, 'epoch': 4.5} + 90%|█████████ | 675/750 [1:10:42<07:25, 5.94s/it] 90%|█████████ | 676/750 [1:10:48<07:19, 5.95s/it] {'loss': 0.1304, 'grad_norm': 0.20279021561145782, 'learning_rate': 9.222752929447538e-05, 'epoch': 4.51} + 90%|█████████ | 676/750 [1:10:48<07:19, 5.95s/it] 90%|█████████ | 677/750 [1:10:54<07:15, 5.96s/it] {'loss': 0.11, 'grad_norm': 0.19248034060001373, 'learning_rate': 9.215938925614141e-05, 'epoch': 4.51} + 90%|█████████ | 677/750 [1:10:54<07:15, 5.96s/it] 90%|█████████ | 678/750 [1:11:00<07:09, 5.96s/it] {'loss': 0.1148, 'grad_norm': 0.19480571150779724, 'learning_rate': 9.209140002574652e-05, 'epoch': 4.52} + 90%|█████████ | 678/750 [1:11:00<07:09, 5.96s/it] 91%|█████████ | 679/750 [1:11:06<07:03, 5.96s/it] {'loss': 0.1315, 'grad_norm': 0.2050219625234604, 'learning_rate': 9.20235610478285e-05, 'epoch': 4.53} + 91%|█████████ | 679/750 [1:11:06<07:03, 5.96s/it] 91%|█████████ | 680/750 [1:11:12<06:56, 5.95s/it] {'loss': 0.1219, 'grad_norm': 0.1935853213071823, 'learning_rate': 9.19558717697852e-05, 'epoch': 4.53} + 91%|█████████ | 680/750 [1:11:12<06:56, 5.95s/it] 91%|█████████ | 681/750 [1:11:18<06:48, 5.92s/it] {'loss': 0.1207, 'grad_norm': 0.2061282992362976, 'learning_rate': 9.18883316418556e-05, 'epoch': 4.54} + 91%|█████████ | 681/750 [1:11:18<06:48, 5.92s/it] 91%|█████████ | 682/750 [1:11:24<06:43, 5.93s/it] {'loss': 0.1143, 'grad_norm': 0.19050821661949158, 'learning_rate': 9.182094011710111e-05, 'epoch': 4.55} + 91%|█████████ | 682/750 [1:11:24<06:43, 5.93s/it] 91%|█████████ | 683/750 [1:11:29<06:36, 5.92s/it] {'loss': 0.124, 'grad_norm': 0.1824985146522522, 'learning_rate': 9.175369665138691e-05, 'epoch': 4.55} + 91%|█████████ | 683/750 [1:11:29<06:36, 5.92s/it] 91%|█████████ | 684/750 [1:11:35<06:31, 5.93s/it] {'loss': 0.1162, 'grad_norm': 0.18643681704998016, 'learning_rate': 9.168660070336348e-05, 'epoch': 4.56} + 91%|█████████ | 684/750 [1:11:35<06:31, 5.93s/it] 91%|█████████▏| 685/750 [1:11:41<06:23, 5.91s/it] {'loss': 0.1196, 'grad_norm': 0.2066197693347931, 'learning_rate': 9.161965173444834e-05, 'epoch': 4.57} + 91%|█████████▏| 685/750 [1:11:41<06:23, 5.91s/it] 91%|█████████▏| 686/750 [1:11:47<06:18, 5.91s/it] {'loss': 0.1396, 'grad_norm': 0.23195287585258484, 'learning_rate': 9.155284920880796e-05, 'epoch': 4.57} + 91%|█████████▏| 686/750 [1:11:47<06:18, 5.91s/it] 92%|█████████▏| 687/750 [1:11:53<06:13, 5.93s/it] {'loss': 0.1106, 'grad_norm': 0.18247000873088837, 'learning_rate': 9.148619259333955e-05, 'epoch': 4.58} + 92%|█████████▏| 687/750 [1:11:53<06:13, 5.93s/it] 92%|█████████▏| 688/750 [1:11:59<06:08, 5.94s/it] {'loss': 0.1161, 'grad_norm': 0.1833108812570572, 'learning_rate': 9.141968135765342e-05, 'epoch': 4.59} + 92%|█████████▏| 688/750 [1:11:59<06:08, 5.94s/it] 92%|█████████▏| 689/750 [1:12:05<06:00, 5.91s/it] {'loss': 0.116, 'grad_norm': 0.19194342195987701, 'learning_rate': 9.135331497405505e-05, 'epoch': 4.59} + 92%|█████████▏| 689/750 [1:12:05<06:00, 5.91s/it] 92%|█████████▏| 690/750 [1:12:11<05:54, 5.90s/it] {'loss': 0.1349, 'grad_norm': 0.20306549966335297, 'learning_rate': 9.128709291752768e-05, 'epoch': 4.6} + 92%|█████████▏| 690/750 [1:12:11<05:54, 5.90s/it] 92%|█████████▏| 691/750 [1:12:17<05:49, 5.92s/it] {'loss': 0.1342, 'grad_norm': 0.19383618235588074, 'learning_rate': 9.122101466571477e-05, 'epoch': 4.61} + 92%|█████████▏| 691/750 [1:12:17<05:49, 5.92s/it] 92%|█████████▏| 692/750 [1:12:23<05:45, 5.96s/it] {'loss': 0.1214, 'grad_norm': 0.18854767084121704, 'learning_rate': 9.115507969890265e-05, 'epoch': 4.61} + 92%|█████████▏| 692/750 [1:12:23<05:45, 5.96s/it] 92%|█████████▏| 693/750 [1:12:29<05:38, 5.95s/it] {'loss': 0.133, 'grad_norm': 0.2005634307861328, 'learning_rate': 9.108928750000352e-05, 'epoch': 4.62} + 92%|█████████▏| 693/750 [1:12:29<05:38, 5.95s/it] 93%|█████████▎| 694/750 [1:12:35<05:31, 5.92s/it] {'loss': 0.1229, 'grad_norm': 0.19461499154567719, 'learning_rate': 9.102363755453824e-05, 'epoch': 4.63} + 93%|█████████▎| 694/750 [1:12:35<05:31, 5.92s/it] 93%|█████████▎| 695/750 [1:12:41<05:26, 5.93s/it] {'loss': 0.1194, 'grad_norm': 0.1921541690826416, 'learning_rate': 9.095812935061951e-05, 'epoch': 4.63} + 93%|█████████▎| 695/750 [1:12:41<05:26, 5.93s/it] 93%|█████████▎| 696/750 [1:12:47<05:20, 5.93s/it] {'loss': 0.1241, 'grad_norm': 0.1985134333372116, 'learning_rate': 9.089276237893512e-05, 'epoch': 4.64} + 93%|█████████▎| 696/750 [1:12:47<05:20, 5.93s/it] 93%|█████████▎| 697/750 [1:12:52<05:13, 5.91s/it] {'loss': 0.1144, 'grad_norm': 0.20523406565189362, 'learning_rate': 9.082753613273131e-05, 'epoch': 4.65} + 93%|█████████▎| 697/750 [1:12:52<05:13, 5.91s/it] 93%|█████████▎| 698/750 [1:12:58<05:06, 5.89s/it] {'loss': 0.1206, 'grad_norm': 0.19906945526599884, 'learning_rate': 9.076245010779625e-05, 'epoch': 4.65} + 93%|█████████▎| 698/750 [1:12:58<05:06, 5.89s/it] 93%|█████████▎| 699/750 [1:13:04<05:01, 5.91s/it] {'loss': 0.1092, 'grad_norm': 0.198943093419075, 'learning_rate': 9.069750380244364e-05, 'epoch': 4.66} + 93%|█████████▎| 699/750 [1:13:04<05:01, 5.91s/it] 93%|█████████▎| 700/750 [1:13:10<04:56, 5.93s/it] {'loss': 0.1142, 'grad_norm': 0.19928747415542603, 'learning_rate': 9.063269671749657e-05, 'epoch': 4.67} + 93%|█████████▎| 700/750 [1:13:10<04:56, 5.93s/it] 93%|█████████▎| 701/750 [1:13:16<04:49, 5.92s/it] {'loss': 0.1122, 'grad_norm': 0.17332346737384796, 'learning_rate': 9.056802835627134e-05, 'epoch': 4.67} + 93%|█████████▎| 701/750 [1:13:16<04:49, 5.92s/it] 94%|█████████▎| 702/750 [1:13:22<04:45, 5.94s/it] {'loss': 0.1274, 'grad_norm': 0.2081272155046463, 'learning_rate': 9.050349822456143e-05, 'epoch': 4.68} + 94%|█████████▎| 702/750 [1:13:22<04:45, 5.94s/it] 94%|█████████▎| 703/750 [1:13:28<04:38, 5.92s/it] {'loss': 0.1178, 'grad_norm': 0.21539074182510376, 'learning_rate': 9.043910583062168e-05, 'epoch': 4.69} + 94%|█████████▎| 703/750 [1:13:28<04:38, 5.92s/it] 94%|█████████▍| 704/750 [1:13:34<04:31, 5.91s/it] {'loss': 0.1394, 'grad_norm': 0.21388739347457886, 'learning_rate': 9.037485068515265e-05, 'epoch': 4.69} + 94%|█████████▍| 704/750 [1:13:34<04:31, 5.91s/it] 94%|█████████▍| 705/750 [1:13:40<04:24, 5.88s/it] {'loss': 0.129, 'grad_norm': 0.2084227055311203, 'learning_rate': 9.03107323012848e-05, 'epoch': 4.7} + 94%|█████████▍| 705/750 [1:13:40<04:24, 5.88s/it] 94%|██████��██▍| 706/750 [1:13:46<04:20, 5.92s/it] {'loss': 0.1305, 'grad_norm': 0.18557515740394592, 'learning_rate': 9.024675019456317e-05, 'epoch': 4.71} + 94%|█████████▍| 706/750 [1:13:46<04:20, 5.92s/it] 94%|█████████▍| 707/750 [1:13:52<04:13, 5.91s/it] {'loss': 0.1167, 'grad_norm': 0.19854147732257843, 'learning_rate': 9.01829038829319e-05, 'epoch': 4.71} + 94%|█████████▍| 707/750 [1:13:52<04:13, 5.91s/it] 94%|█████████▍| 708/750 [1:13:57<04:07, 5.90s/it] {'loss': 0.1169, 'grad_norm': 0.18760545551776886, 'learning_rate': 9.011919288671903e-05, 'epoch': 4.72} + 94%|█████████▍| 708/750 [1:13:57<04:07, 5.90s/it] 95%|█████████▍| 709/750 [1:14:03<04:01, 5.90s/it] {'loss': 0.1275, 'grad_norm': 0.17979490756988525, 'learning_rate': 9.005561672862141e-05, 'epoch': 4.73} + 95%|█████████▍| 709/750 [1:14:03<04:01, 5.90s/it] 95%|█████████▍| 710/750 [1:14:09<03:56, 5.92s/it] {'loss': 0.1218, 'grad_norm': 0.1948082447052002, 'learning_rate': 8.999217493368951e-05, 'epoch': 4.73} + 95%|█████████▍| 710/750 [1:14:09<03:56, 5.92s/it] 95%|█████████▍| 711/750 [1:14:15<03:51, 5.95s/it] {'loss': 0.133, 'grad_norm': 0.20126453042030334, 'learning_rate': 8.992886702931274e-05, 'epoch': 4.74} + 95%|█████████▍| 711/750 [1:14:15<03:51, 5.95s/it] 95%|█████████▍| 712/750 [1:14:21<03:46, 5.97s/it] {'loss': 0.1281, 'grad_norm': 0.19406697154045105, 'learning_rate': 8.986569254520445e-05, 'epoch': 4.75} + 95%|█████████▍| 712/750 [1:14:21<03:46, 5.97s/it] 95%|█████████▌| 713/750 [1:14:27<03:41, 5.97s/it] {'loss': 0.1313, 'grad_norm': 0.1871439665555954, 'learning_rate': 8.980265101338746e-05, 'epoch': 4.75} + 95%|█████████▌| 713/750 [1:14:27<03:41, 5.97s/it] 95%|█████████▌| 714/750 [1:14:33<03:34, 5.96s/it] {'loss': 0.1083, 'grad_norm': 0.17861828207969666, 'learning_rate': 8.973974196817932e-05, 'epoch': 4.76} + 95%|█████████▌| 714/750 [1:14:33<03:34, 5.96s/it] 95%|█████████▌| 715/750 [1:14:39<03:28, 5.96s/it] {'loss': 0.1126, 'grad_norm': 0.18450631201267242, 'learning_rate': 8.967696494617803e-05, 'epoch': 4.77} + 95%|█████████▌| 715/750 [1:14:39<03:28, 5.96s/it] 95%|█████████▌| 716/750 [1:14:45<03:22, 5.96s/it] {'loss': 0.1183, 'grad_norm': 0.18953324854373932, 'learning_rate': 8.961431948624752e-05, 'epoch': 4.77} + 95%|█████████▌| 716/750 [1:14:45<03:22, 5.96s/it] 96%|█████████▌| 717/750 [1:14:51<03:16, 5.96s/it] {'loss': 0.1129, 'grad_norm': 0.17921291291713715, 'learning_rate': 8.955180512950356e-05, 'epoch': 4.78} + 96%|█████████▌| 717/750 [1:14:51<03:16, 5.96s/it] 96%|█████████▌| 718/750 [1:14:57<03:10, 5.97s/it] {'loss': 0.108, 'grad_norm': 0.18071863055229187, 'learning_rate': 8.948942141929963e-05, 'epoch': 4.79} + 96%|█████████▌| 718/750 [1:14:57<03:10, 5.97s/it] 96%|█████████▌| 719/750 [1:15:03<03:05, 5.97s/it] {'loss': 0.1286, 'grad_norm': 0.20430976152420044, 'learning_rate': 8.942716790121283e-05, 'epoch': 4.79} + 96%|█████████▌| 719/750 [1:15:03<03:05, 5.97s/it] 96%|█████████▌| 720/750 [1:15:09<02:59, 5.97s/it] {'loss': 0.1151, 'grad_norm': 0.18577136099338531, 'learning_rate': 8.936504412303008e-05, 'epoch': 4.8} + 96%|█████████▌| 720/750 [1:15:09<02:59, 5.97s/it] 96%|█████████▌| 721/750 [1:15:15<02:53, 5.98s/it] {'loss': 0.1387, 'grad_norm': 0.19394683837890625, 'learning_rate': 8.93030496347342e-05, 'epoch': 4.81} + 96%|█████████▌| 721/750 [1:15:15<02:53, 5.98s/it] 96%|█████████▋| 722/750 [1:15:21<02:47, 5.98s/it] {'loss': 0.1254, 'grad_norm': 0.19409403204917908, 'learning_rate': 8.924118398849037e-05, 'epoch': 4.81} + 96%|█████████▋| 722/750 [1:15:21<02:47, 5.98s/it] 96%|█████████▋| 723/750 [1:15:27<02:41, 5.97s/it] {'loss': 0.15, 'grad_norm': 0.19994400441646576, 'learning_rate': 8.917944673863242e-05, 'epoch': 4.82} + 96%|█████████▋| 723/750 [1:15:27<02:41, 5.97s/it] 97%|█████████▋| 724/750 [1:15:33<02:34, 5.95s/it] {'loss': 0.1187, 'grad_norm': 0.19855041801929474, 'learning_rate': 8.911783744164937e-05, 'epoch': 4.83} + 97%|█████████▋| 724/750 [1:15:33<02:34, 5.95s/it] 97%|█████████▋| 725/750 [1:15:39<02:28, 5.94s/it] {'loss': 0.12, 'grad_norm': 0.20143508911132812, 'learning_rate': 8.905635565617214e-05, 'epoch': 4.83} + 97%|█████████▋| 725/750 [1:15:39<02:28, 5.94s/it] 97%|█████████▋| 726/750 [1:15:45<02:22, 5.94s/it] {'loss': 0.1274, 'grad_norm': 0.196522518992424, 'learning_rate': 8.899500094296009e-05, 'epoch': 4.84} + 97%|█████████▋| 726/750 [1:15:45<02:22, 5.94s/it] 97%|█████████▋| 727/750 [1:15:51<02:16, 5.95s/it] {'loss': 0.1161, 'grad_norm': 0.18777668476104736, 'learning_rate': 8.893377286488804e-05, 'epoch': 4.85} + 97%|█████████▋| 727/750 [1:15:51<02:16, 5.95s/it] 97%|█████████▋| 728/750 [1:15:57<02:12, 6.00s/it] {'loss': 0.1204, 'grad_norm': 0.1800437867641449, 'learning_rate': 8.887267098693303e-05, 'epoch': 4.85} + 97%|█████████▋| 728/750 [1:15:57<02:12, 6.00s/it] 97%|█████████▋| 729/750 [1:16:03<02:05, 5.99s/it] {'loss': 0.1282, 'grad_norm': 0.20204044878482819, 'learning_rate': 8.881169487616148e-05, 'epoch': 4.86} + 97%|█████████▋| 729/750 [1:16:03<02:05, 5.99s/it] 97%|█████████▋| 730/750 [1:16:09<01:59, 5.96s/it] {'loss': 0.1117, 'grad_norm': 0.20082896947860718, 'learning_rate': 8.875084410171615e-05, 'epoch': 4.87} + 97%|█████████▋| 730/750 [1:16:09<01:59, 5.96s/it] 97%|█████████▋| 731/750 [1:16:15<01:53, 5.98s/it] {'loss': 0.1343, 'grad_norm': 0.19596554338932037, 'learning_rate': 8.869011823480348e-05, 'epoch': 4.87} + 97%|█████████▋| 731/750 [1:16:15<01:53, 5.98s/it] 98%|█████████▊| 732/750 [1:16:21<01:47, 5.97s/it] {'loss': 0.1305, 'grad_norm': 0.21354715526103973, 'learning_rate': 8.862951684868085e-05, 'epoch': 4.88} + 98%|█████████▊| 732/750 [1:16:21<01:47, 5.97s/it] 98%|█████████▊| 733/750 [1:16:27<01:41, 5.96s/it] {'loss': 0.1263, 'grad_norm': 0.1988142877817154, 'learning_rate': 8.856903951864397e-05, 'epoch': 4.89} + 98%|█████████▊| 733/750 [1:16:27<01:41, 5.96s/it] 98%|█████████▊| 734/750 [1:16:33<01:35, 5.98s/it] {'loss': 0.1193, 'grad_norm': 0.19559569656848907, 'learning_rate': 8.85086858220144e-05, 'epoch': 4.89} + 98%|█████████▊| 734/750 [1:16:33<01:35, 5.98s/it] 98%|█████████▊| 735/750 [1:16:39<01:29, 5.97s/it] {'loss': 0.1301, 'grad_norm': 0.20791786909103394, 'learning_rate': 8.844845533812719e-05, 'epoch': 4.9} + 98%|█████████▊| 735/750 [1:16:39<01:29, 5.97s/it] 98%|█████████▊| 736/750 [1:16:44<01:23, 5.95s/it] {'loss': 0.1327, 'grad_norm': 0.20150598883628845, 'learning_rate': 8.838834764831844e-05, 'epoch': 4.91} + 98%|█████████▊| 736/750 [1:16:44<01:23, 5.95s/it] 98%|█████████▊| 737/750 [1:16:50<01:17, 5.94s/it] {'loss': 0.1253, 'grad_norm': 0.1925792694091797, 'learning_rate': 8.83283623359132e-05, 'epoch': 4.91} + 98%|█████████▊| 737/750 [1:16:50<01:17, 5.94s/it] 98%|█████████▊| 738/750 [1:16:56<01:11, 5.92s/it] {'loss': 0.1123, 'grad_norm': 0.19202423095703125, 'learning_rate': 8.826849898621327e-05, 'epoch': 4.92} + 98%|█████████▊| 738/750 [1:16:56<01:11, 5.92s/it] 99%|█████████▊| 739/750 [1:17:02<01:05, 5.92s/it] {'loss': 0.1194, 'grad_norm': 0.21489877998828888, 'learning_rate': 8.82087571864852e-05, 'epoch': 4.93} + 99%|█████���███▊| 739/750 [1:17:02<01:05, 5.92s/it] 99%|█████████▊| 740/750 [1:17:08<00:59, 5.90s/it] {'loss': 0.1247, 'grad_norm': 0.19145846366882324, 'learning_rate': 8.814913652594829e-05, 'epoch': 4.93} + 99%|█████████▊| 740/750 [1:17:08<00:59, 5.90s/it] 99%|█████████▉| 741/750 [1:17:14<00:52, 5.88s/it] {'loss': 0.14, 'grad_norm': 0.21258756518363953, 'learning_rate': 8.808963659576277e-05, 'epoch': 4.94} + 99%|█████████▉| 741/750 [1:17:14<00:52, 5.88s/it] 99%|█████████▉| 742/750 [1:17:20<00:46, 5.87s/it] {'loss': 0.1338, 'grad_norm': 0.20081067085266113, 'learning_rate': 8.803025698901805e-05, 'epoch': 4.95} + 99%|█████████▉| 742/750 [1:17:20<00:46, 5.87s/it] 99%|█████████▉| 743/750 [1:17:26<00:41, 5.89s/it] {'loss': 0.1243, 'grad_norm': 0.20036664605140686, 'learning_rate': 8.797099730072091e-05, 'epoch': 4.95} + 99%|█████████▉| 743/750 [1:17:26<00:41, 5.89s/it] 99%|█████████▉| 744/750 [1:17:31<00:35, 5.88s/it] {'loss': 0.1208, 'grad_norm': 0.19513240456581116, 'learning_rate': 8.791185712778405e-05, 'epoch': 4.96} + 99%|█████████▉| 744/750 [1:17:32<00:35, 5.88s/it] 99%|█████████▉| 745/750 [1:17:37<00:29, 5.88s/it] {'loss': 0.1249, 'grad_norm': 0.2070721685886383, 'learning_rate': 8.785283606901446e-05, 'epoch': 4.97} + 99%|█████████▉| 745/750 [1:17:37<00:29, 5.88s/it] 99%|█████████▉| 746/750 [1:17:43<00:23, 5.87s/it] {'loss': 0.152, 'grad_norm': 0.2183961719274521, 'learning_rate': 8.779393372510207e-05, 'epoch': 4.97} + 99%|█████████▉| 746/750 [1:17:43<00:23, 5.87s/it] 100%|█████████▉| 747/750 [1:17:49<00:17, 5.92s/it] {'loss': 0.1179, 'grad_norm': 0.20051303505897522, 'learning_rate': 8.773514969860834e-05, 'epoch': 4.98} + 100%|█████████▉| 747/750 [1:17:49<00:17, 5.92s/it] 100%|█████████▉| 748/750 [1:17:55<00:11, 5.93s/it] {'loss': 0.128, 'grad_norm': 0.18662500381469727, 'learning_rate': 8.767648359395506e-05, 'epoch': 4.99} + 100%|█████████▉| 748/750 [1:17:55<00:11, 5.93s/it] 100%|█████████▉| 749/750 [1:18:01<00:05, 5.96s/it] {'loss': 0.1166, 'grad_norm': 0.18965964019298553, 'learning_rate': 8.761793501741308e-05, 'epoch': 4.99} + 100%|█████████▉| 749/750 [1:18:01<00:05, 5.96s/it] 100%|██████████| 750/750 [1:18:09<00:00, 6.58s/it] {'loss': 0.1321, 'grad_norm': 0.2086346298456192, 'learning_rate': 8.755950357709131e-05, 'epoch': 5.0} + 100%|██████████| 750/750 [1:18:09<00:00, 6.58s/it][INFO|trainer.py:3515] 2025-06-27 01:07:06,408 >> Saving model checkpoint to ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-750 +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da162-7a76ce6814b2933b0506bdc8;e31c84cb-902c-45c7-a8c4-68586d16784b) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +[INFO|tokenization_utils_base.py:2684] 2025-06-27 01:07:06,802 >> tokenizer config file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-750/tokenizer_config.json +[INFO|tokenization_utils_base.py:2693] 2025-06-27 01:07:06,803 >> Special tokens file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-750/special_tokens_map.json +[2025-06-27 01:07:11,102] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step750 is begin to save! +[2025-06-27 01:07:11,124] [INFO] [logging.py:107:log_dist] [Rank 0] Saving model checkpoint: ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-750/global_step750/mp_rank_00_model_states.pt +[INFO|trainer.py:3607] 2025-06-27 01:07:11,253 >> Deleting older checkpoint [outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-600] due to args.save_total_limit +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da167-0da98d195b602ddd1da097f1;5b85b711-8b7f-4848-99fb-8db60242d425) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da167-171c83702e81a00c50282adb;20a3a0db-c780-4f86-a174-de59c54540bd) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da167-401573c35c22a55a47e63ebe;8f6274b5-a237-441b-9042-e65a0e5353e5) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +[INFO|trainer.py:2406] 2025-06-27 01:07:11,646 >> + +Training completed. Do not forget to share your model on huggingface.co/models =) + + +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da167-71037e814a30726033b63379;07ef2bfb-df7d-4317-a41d-e71761b82666) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +[INFO|trainer.py:2644] 2025-06-27 01:07:11,978 >> Loading best model from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400 (score: 0.20738115906715393). +[INFO|deepspeed.py:431] 2025-06-27 01:07:11,979 >> Attempting to resume from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400 +[2025-06-27 01:07:11,980] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Begin Load checkpoint from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/global_step400/mp_rank_00_model_states.pt... +[2025-06-27 01:07:11,999] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] End Load checkpoint from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/global_step400/mp_rank_00_model_states.pt... +[2025-06-27 01:07:12,000] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Begin Load checkpoint from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/global_step400/mp_rank_00_model_states.pt... +[2025-06-27 01:07:12,014] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] End Load checkpoint from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/global_step400/mp_rank_00_model_states.pt... +[2025-06-27 01:07:12,045] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Begin Load checkpoint from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/global_step400/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt... +[2025-06-27 01:07:12,060] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] End Load checkpoint from ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-400/global_step400/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt... +[2025-06-27 01:07:12,060] [INFO] [engine.py:3277:_get_all_zero_checkpoint_state_dicts] successfully read 4 ZeRO state_dicts for rank 0 +[2025-06-27 01:07:12,069] [INFO] [engine.py:3227:_load_zero_checkpoint] loading 4 zero partition checkpoints for rank 0 + {'train_runtime': 4707.1456, 'train_samples_per_second': 15.933, 'train_steps_per_second': 0.159, 'train_loss': 0.17488925837477048, 'epoch': 5.0} + 100%|██████████| 750/750 [1:18:25<00:00, 6.58s/it][INFO|trainer.py:2447] 2025-06-27 01:07:12,078 >> Deleting older checkpoint [outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/checkpoint-750] due to args.save_total_limit +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da168-19b7bd84391e677018665b40;8110803f-ebfc-4ef4-a01c-3cbb7ad128b2) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da168-2de248032402e78410e9eb77;abd1291e-07b4-4651-9d36-031537ee247a) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da168-5add01581039cf7922a03e78;86e3433b-5446-4523-a1e9-5a8830faaedd) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da168-4623087c69fb423053ad278d;5951f658-6b53-4066-8b5c-8d199dfd3509) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( + 100%|██████████| 750/750 [1:18:25<00:00, 6.27s/it] +[INFO|trainer.py:3515] 2025-06-27 01:07:21,803 >> Saving model checkpoint to ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/ +/home/iitm_admin/miniconda3/envs/midalign/lib/python3.9/site-packages/peft/utils/other.py:1110: UserWarning: Unable to fetch remote file due to the following error 403 Client Error. (Request ID: Root=1-685da171-649b365b75223cc009186264;2de77f84-2317-42a3-9e2b-4685dd5ba448) + +Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json. +Access to model meta-llama/Llama-3.1-8B-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct to ask for access. - silently ignoring the lookup for the file config.json in meta-llama/Llama-3.1-8B-Instruct. + warnings.warn( +[INFO|tokenization_utils_base.py:2684] 2025-06-27 01:07:22,131 >> tokenizer config file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/tokenizer_config.json +[INFO|tokenization_utils_base.py:2693] 2025-06-27 01:07:22,132 >> Special tokens file saved in ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/special_tokens_map.json +***** train metrics ***** + epoch = 5.0 + total_flos = 6234656320GF + train_loss = 0.1749 + train_runtime = 1:18:27.14 + train_samples = 15000 + train_samples_per_second = 15.933 + train_steps_per_second = 0.159 +06/27/2025 01:07:23 - INFO - __main__ - *** Evaluate *** +[INFO|trainer.py:3831] 2025-06-27 01:07:23,242 >> +***** Running Evaluation ***** +[INFO|trainer.py:3833] 2025-06-27 01:07:23,242 >> Num examples = 1000 +[INFO|trainer.py:3836] 2025-06-27 01:07:23,242 >> Batch size = 25 + 0%| | 0/10 [00:00> Dropping the following result as it does not have all the necessary fields: +{'task': {'name': 'Causal Language Modeling', 'type': 'text-generation'}, 'metrics': [{'name': 'Accuracy', 'type': 'accuracy', 'value': 0.22317733268197362}]} +wandb: +wandb: 🚀 View run ./outputs/data/reason/meta-llama/Llama-3.1-8B-Instruct/bn/baseline/data_15000_1000/ at: https://wandb.ai/indic-encoder/midalign/runs/s645lnzf +wandb: Find logs at: wandb/run-20250626_234845-s645lnzf/logs +[rank0]:[W627 01:08:06.152459494 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())