| architecture: |
| training_mode: lora |
| gradient_checkpointing: true |
| intermediate_dropout: 0.0 |
| backbone_kwargs: '{}' |
| dataset: |
| data_sample_choice: |
| - Train |
| - Validation |
| data_sample: 0.2 |
| system_column: null |
| input_column: prompt |
| output_column: answer |
| train_dataset_id: e3a038d6-c009-4935-97a3-d35720bee1de |
| train_dataframe: http://minio:9000/datasets/e3a038d6-c009-4935-97a3-d35720bee1de/e3a038d6-c009-4935-97a3-d35720bee1de?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=minioadmin%2F20251017%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251017T161637Z&X-Amz-Expires=604800&X-Amz-SignedHeaders=host&X-Amz-Signature=181921582822dc2618a04e8dc3a9978aa6d26de465d977b024400311fb77376d |
| train_dataframe_type: parquet |
| validation_dataframe: null |
| validation_dataframe_type: null |
| validation_strategy: automatic |
| validation_size: 0.2 |
| num_classes: null |
| validation_dataset_id: null |
| unroll_conversations: false |
| environment: |
| find_unused_parameters: false |
| huggingface_branch: main |
| mixed_precision: false |
| use_fsdp: false |
| use_fsdp_cpu_offload: false |
| seed: -1 |
| trust_remote_code: true |
| tokenizer: |
| max_length: 2304 |
| padding_quantile: 1.0 |
| tokenizer_kwargs: '{"use_fast": true, "add_prefix_space": false}' |
| padding_side: left |
| chat_template: null |
| training: |
| attention_implementation: auto |
| batch_size: 1 |
| differential_learning_rate_layers: [] |
| differential_learning_rate: 1.0e-05 |
| epochs: 1 |
| evaluate_before_training: true |
| evaluation_epochs: 1.0 |
| grad_accumulation: 1 |
| gradient_clip: 0.0 |
| learning_rate: 0.0001 |
| lora_alpha: 16 |
| lora_dropout: 0.05 |
| lora_rank: 4 |
| lora_target_modules: '' |
| loss_function: CrossEntropyLoss |
| min_learning_rate_ratio: 0.0 |
| optimizer: AdamW |
| schedule: Cosine |
| train_validation_data: false |
| use_length_based_sampler: false |
| warmup_epochs: 0.0 |
| weight_decay: 0.0 |
| llm_judge_model: '' |
| llm_judge_prompt_template: '' |
| logging: {} |
| prediction: |
| batch_size_inference: 0 |
| max_length_inference: 256 |
| metrics: |
| - BLEU |
| min_length_inference: 2 |
| num_beams: 1 |
| repetition_penalty: 1.0 |
| temperature: 0.0 |
| top_k: 0 |
| top_p: 1.0 |
| experiment_name: hungry-bonobo |
| experiment_id: 74273746-83a7-459b-83ed-c4d06df0d782 |
| project_id: 14673bfa-20ad-486c-815b-f0fed6c6eae8 |
| reference_experiment_ids: [] |
| llm_backbone: h2oai/h2ovl-mississippi-2b |
| problem_type: multimodal_causal_language_modeling |
|
|