| dataset: |
| align_stage_components: |
| - download/med-v0.1-instruct/med_v0_1_mix.json |
| - download/med-v0.1-instruct |
| dataset_id: med-instruct |
| dataset_root_dir: data |
| finetune_stage_components: |
| - download/med-v0.1-instruct/med_v0_1_mix.json |
| - download/med-v0.1-instruct |
| type: med-instruct |
| model: |
| align_epochs: 1 |
| align_global_batch_size: 256 |
| align_learning_rate: 0.001 |
| align_lr_scheduler_type: linear-warmup+cosine-decay |
| align_max_grad_norm: 1.0 |
| align_max_steps: null |
| align_per_device_batch_size: 16 |
| align_train_strategy: fsdp-shard-grad-op |
| align_warmup_ratio: 0.03 |
| align_weight_decay: 0.0 |
| arch_specifier: no-align+gelu-mlp |
| enable_gradient_checkpointing: true |
| enable_mixed_precision_training: true |
| finetune_epochs: 5 |
| finetune_global_batch_size: 128 |
| finetune_learning_rate: 2.0e-06 |
| finetune_lr_scheduler_type: linear-warmup+cosine-decay |
| finetune_max_grad_norm: 1.0 |
| finetune_max_steps: null |
| finetune_per_device_batch_size: 16 |
| finetune_train_strategy: fsdp-full-shard |
| finetune_warmup_ratio: 0.03 |
| finetune_weight_decay: 0.1 |
| image_resize_strategy: letterbox |
| llm_backbone_id: llama2-7b-chat |
| llm_max_length: 2048 |
| model_id: testmodel |
| reduce_in_full_precision: false |
| type: one-stage+7b |
| vision_backbone_id: clip-vit-l-336px |
| pretrained_checkpoint: null |
| run_id: med-instruct+testmodel+stage-finetune+x7 |
| run_root_dir: runs |
| seed: 7 |
| stage: finetune |
| trackers: [] |
| wandb_entity: stanford-voltron |
| wandb_project: onyx-vlms |
|
|