| ? '' |
| : ? '' |
| : ? '' |
| : hydra: |
| run: |
| dir: . |
| output_subdir: null |
| job: |
| chdir: false |
| _target_: null |
| job: |
| name: ??? |
| mode: slurm |
| work_dir: null |
| dry_run: false |
| slurm: |
| time_limit: ??? |
| constraint: |
| - h200 |
| memory: 200 |
| cpus_per_task: 16 |
| partition: null |
| mail_user: user@example.com |
| execution: |
| nodes: null |
| gpus_per_node: null |
| num_gpus: null |
| hostfile: null |
| secrets_file: null |
| model: |
| name_or_path: ??? |
| finetuning_type: lora |
| dataset: |
| name: ??? |
| dir: null |
| info_json: null |
| template: default |
| cutoff_len: 1024 |
| val_size: 0.1 |
| hf_hub_url: null |
| formatting: alpaca |
| ranking: false |
| subset: null |
| split: train |
| folder: null |
| num_samples: null |
| columns: |
| prompt: null |
| query: null |
| response: null |
| history: null |
| messages: null |
| system: null |
| tools: null |
| images: null |
| videos: null |
| audios: null |
| chosen: null |
| rejected: null |
| kto_tag: null |
| tags: |
| role: null |
| content: null |
| user: null |
| assistant: null |
| observation: null |
| function: null |
| system: null |
| training: |
| stage: sft |
| do_train: true |
| model_name_or_path: null |
| finetuning_type: lora |
| trust_remote_code: true |
| dataset: null |
| dataset_dir: null |
| template: default |
| cutoff_len: 1024 |
| val_size: 0.1 |
| preprocessing_num_workers: 1 |
| dataset_num_proc: 1 |
| dataloader_num_workers: 0 |
| streaming: false |
| learning_rate: 5.0e-05 |
| num_train_epochs: 3.0 |
| per_device_train_batch_size: 1 |
| per_device_eval_batch_size: 1 |
| gradient_accumulation_steps: 8 |
| lr_scheduler_type: cosine |
| warmup_ratio: 0.1 |
| warmup_steps: 0 |
| lora_rank: 8 |
| lora_alpha: 16 |
| lora_dropout: 0.05 |
| lora_target: all |
| optim: adamw_torch |
| bf16: true |
| fp16: false |
| output_dir: null |
| save_strategy: epoch |
| save_steps: 500 |
| save_total_limit: 3 |
| save_only_model: false |
| eval_strategy: steps |
| eval_steps: 500 |
| do_eval: true |
| logging_steps: 10 |
| plot_loss: true |
| report_to: none |
| gradient_checkpointing: true |
| ddp_timeout: 180000000 |
| include_num_input_tokens_seen: true |
| overwrite_output_dir: true |
| overwrite_cache: false |
| seed: 42 |
| lora: |
| rank: 8 |
| alpha: 16 |
| dropout: 0.05 |
| target: all |
| output: |
| experiment_dir: ./experiments |
| merge: |
| stage: export |
| model_name_or_path: null |
| adapter_name_or_path: null |
| template: default |
| export_dir: null |
| export_size: 2 |
| export_device: auto |
| export_legacy_format: false |
| finetuning_type: lora |
| wandb: |
| project: null |
| run_name: null |
| entity: null |
| hf: |
| repo_id: null |
| private: false |
| upload_artifacts: true |
| cleanup: |
| checkpoints: false |
| merged: false |
| job: |
| name: lf_torch_test__interactive |
| mode: local |
| work_dir: null |
| dry_run: false |
| slurm: |
| time_limit: null |
| constraint: null |
| memory: null |
| partition: null |
| mail_user: null |
| execution: |
| nodes: 2 |
| gpus_per_node: 2 |
| num_gpus: null |
| hostfile: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/hostfile_auto_generated.txt |
| secrets_file: ./secrets.env |
| model: |
| name_or_path: Qwen/Qwen2.5-0.5B |
| finetuning_type: lora |
| lora: |
| rank: 8 |
| alpha: 16 |
| dropout: 0.05 |
| target: all |
| dataset: |
| name: my_custom_sft12 |
| dir: null |
| info_json: null |
| template: default |
| cutoff_len: 8096 |
| val_size: 0.1 |
| hf_hub_url: TAUR-dev/D-SFT_C-BASELINE_r1_distillation-sft-data |
| formatting: sharegpt |
| ranking: false |
| subset: null |
| split: train |
| folder: null |
| num_samples: null |
| columns: |
| messages: conversations |
| tags: |
| role: role |
| content: content |
| user: user |
| assistant: assistant |
| tokenized_path: /scratch/zrs2020/LlamaFactoryHelper/experiments/lf_torch_test__interactive/tokenized/my_custom_sft12 |
| data_shared_file_system: true |
| output: |
| experiment_dir: ./experiments |
| wandb: |
| project: null |
| run_name: interactive_test |
| entity: null |
| hf: |
| repo_id: TAUR-dev/testing_llamafactory_helper_quick_test__interactive |
| private: false |
| cleanup: |
| checkpoints: false |
| merged: false |
| training: |
| stage: sft |
| do_train: true |
| max_steps: 100 |
| do_eval: false |
| save_strategy: steps |
| save_steps: 50 |
| logging_steps: 10 |
| fp16: true |
| bf16: false |
| overwrite_output_dir: true |
| per_device_train_batch_size: 1 |
| gradient_accumulation_steps: 1 |
| gradient_checkpointing: true |
| preprocessing_num_workers: 16 |
| overwrite_cache: true |
| merge: {} |
|
|