test5 / cfg.yaml
guanshuo's picture
Pushing model from H2O LLM Studio - Experiment 74273746-83a7-459b-83ed-c4d06df0d782
f209c11 verified
architecture:
training_mode: lora
gradient_checkpointing: true
intermediate_dropout: 0.0
backbone_kwargs: '{}'
dataset:
data_sample_choice:
- Train
- Validation
data_sample: 0.2
system_column: null
input_column: prompt
output_column: answer
train_dataset_id: e3a038d6-c009-4935-97a3-d35720bee1de
train_dataframe: http://minio:9000/datasets/e3a038d6-c009-4935-97a3-d35720bee1de/e3a038d6-c009-4935-97a3-d35720bee1de?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=minioadmin%2F20251017%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251017T161637Z&X-Amz-Expires=604800&X-Amz-SignedHeaders=host&X-Amz-Signature=181921582822dc2618a04e8dc3a9978aa6d26de465d977b024400311fb77376d
train_dataframe_type: parquet
validation_dataframe: null
validation_dataframe_type: null
validation_strategy: automatic
validation_size: 0.2
num_classes: null
validation_dataset_id: null
unroll_conversations: false
environment:
find_unused_parameters: false
huggingface_branch: main
mixed_precision: false
use_fsdp: false
use_fsdp_cpu_offload: false
seed: -1
trust_remote_code: true
tokenizer:
max_length: 2304
padding_quantile: 1.0
tokenizer_kwargs: '{"use_fast": true, "add_prefix_space": false}'
padding_side: left
chat_template: null
training:
attention_implementation: auto
batch_size: 1
differential_learning_rate_layers: []
differential_learning_rate: 1.0e-05
epochs: 1
evaluate_before_training: true
evaluation_epochs: 1.0
grad_accumulation: 1
gradient_clip: 0.0
learning_rate: 0.0001
lora_alpha: 16
lora_dropout: 0.05
lora_rank: 4
lora_target_modules: ''
loss_function: CrossEntropyLoss
min_learning_rate_ratio: 0.0
optimizer: AdamW
schedule: Cosine
train_validation_data: false
use_length_based_sampler: false
warmup_epochs: 0.0
weight_decay: 0.0
llm_judge_model: ''
llm_judge_prompt_template: ''
logging: {}
prediction:
batch_size_inference: 0
max_length_inference: 256
metrics:
- BLEU
min_length_inference: 2
num_beams: 1
repetition_penalty: 1.0
temperature: 0.0
top_k: 0
top_p: 1.0
experiment_name: hungry-bonobo
experiment_id: 74273746-83a7-459b-83ed-c4d06df0d782
project_id: 14673bfa-20ad-486c-815b-f0fed6c6eae8
reference_experiment_ids: []
llm_backbone: h2oai/h2ovl-mississippi-2b
problem_type: multimodal_causal_language_modeling