PEFT
Safetensors
model1 / training_config.yaml
magicsquares137's picture
Upload training_config.yaml with huggingface_hub
0206007 verified
adapter: lora
base_model: HuggingFaceTB/SmolLM2-135M
bf16: auto
dataset_prepared_path: ./last_run_prepared
datasets:
- path: mhenrichsen/alpaca_2k_test
type: alpaca
flash_attention: true
gradient_accumulation_steps: 2
gradient_checkpointing: false
learning_rate: 0.0003
logging_steps: 10
lora_alpha: 16
lora_dropout: 0.05
lora_r: 8
lora_target_modules:
- q_proj
- v_proj
max_steps: 100
micro_batch_size: 2
optimizer: adamw_torch
output_dir: ./outputs/admin_20251203_195913
saves_per_epoch: 1
sequence_len: 256
special_tokens:
pad_token: <|endoftext|>
val_set_size: 0.05