guanshuo commited on
Commit
f209c11
·
verified ·
1 Parent(s): be8ab4b

Pushing model from H2O LLM Studio - Experiment 74273746-83a7-459b-83ed-c4d06df0d782

Browse files
Files changed (1) hide show
  1. cfg.yaml +80 -0
cfg.yaml ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ architecture:
2
+ training_mode: lora
3
+ gradient_checkpointing: true
4
+ intermediate_dropout: 0.0
5
+ backbone_kwargs: '{}'
6
+ dataset:
7
+ data_sample_choice:
8
+ - Train
9
+ - Validation
10
+ data_sample: 0.2
11
+ system_column: null
12
+ input_column: prompt
13
+ output_column: answer
14
+ train_dataset_id: e3a038d6-c009-4935-97a3-d35720bee1de
15
+ train_dataframe: http://minio:9000/datasets/e3a038d6-c009-4935-97a3-d35720bee1de/e3a038d6-c009-4935-97a3-d35720bee1de?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=minioadmin%2F20251017%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251017T161637Z&X-Amz-Expires=604800&X-Amz-SignedHeaders=host&X-Amz-Signature=181921582822dc2618a04e8dc3a9978aa6d26de465d977b024400311fb77376d
16
+ train_dataframe_type: parquet
17
+ validation_dataframe: null
18
+ validation_dataframe_type: null
19
+ validation_strategy: automatic
20
+ validation_size: 0.2
21
+ num_classes: null
22
+ validation_dataset_id: null
23
+ unroll_conversations: false
24
+ environment:
25
+ find_unused_parameters: false
26
+ huggingface_branch: main
27
+ mixed_precision: false
28
+ use_fsdp: false
29
+ use_fsdp_cpu_offload: false
30
+ seed: -1
31
+ trust_remote_code: true
32
+ tokenizer:
33
+ max_length: 2304
34
+ padding_quantile: 1.0
35
+ tokenizer_kwargs: '{"use_fast": true, "add_prefix_space": false}'
36
+ padding_side: left
37
+ chat_template: null
38
+ training:
39
+ attention_implementation: auto
40
+ batch_size: 1
41
+ differential_learning_rate_layers: []
42
+ differential_learning_rate: 1.0e-05
43
+ epochs: 1
44
+ evaluate_before_training: true
45
+ evaluation_epochs: 1.0
46
+ grad_accumulation: 1
47
+ gradient_clip: 0.0
48
+ learning_rate: 0.0001
49
+ lora_alpha: 16
50
+ lora_dropout: 0.05
51
+ lora_rank: 4
52
+ lora_target_modules: ''
53
+ loss_function: CrossEntropyLoss
54
+ min_learning_rate_ratio: 0.0
55
+ optimizer: AdamW
56
+ schedule: Cosine
57
+ train_validation_data: false
58
+ use_length_based_sampler: false
59
+ warmup_epochs: 0.0
60
+ weight_decay: 0.0
61
+ llm_judge_model: ''
62
+ llm_judge_prompt_template: ''
63
+ logging: {}
64
+ prediction:
65
+ batch_size_inference: 0
66
+ max_length_inference: 256
67
+ metrics:
68
+ - BLEU
69
+ min_length_inference: 2
70
+ num_beams: 1
71
+ repetition_penalty: 1.0
72
+ temperature: 0.0
73
+ top_k: 0
74
+ top_p: 1.0
75
+ experiment_name: hungry-bonobo
76
+ experiment_id: 74273746-83a7-459b-83ed-c4d06df0d782
77
+ project_id: 14673bfa-20ad-486c-815b-f0fed6c6eae8
78
+ reference_experiment_ids: []
79
+ llm_backbone: h2oai/h2ovl-mississippi-2b
80
+ problem_type: multimodal_causal_language_modeling