| { | |
| "model": "logos29-gemma2_9b", | |
| "family": "gemma2_9b", | |
| "family_name": "Gemma 2 9B", | |
| "base_model": "google/gemma-2-9b-it", | |
| "base_model_quantized": "unsloth/gemma-2-9b-it-bnb-4bit", | |
| "method": "QLoRA (4-bit NF4 + LoRA)", | |
| "framework": "unsloth", | |
| "lora_rank": 16, | |
| "lora_alpha": 16, | |
| "lora_target_modules": [ | |
| "q_proj", | |
| "k_proj", | |
| "v_proj", | |
| "o_proj", | |
| "gate_proj", | |
| "up_proj", | |
| "down_proj" | |
| ], | |
| "epochs": 3, | |
| "effective_batch_size": 8, | |
| "load_in_4bit": true, | |
| "learning_rate": 0.0002, | |
| "lr_scheduler": "cosine", | |
| "max_seq_length": 2048, | |
| "dataset": "logos29_gemma9b.jsonl", | |
| "dataset_size": 1026, | |
| "train_on_responses_only": true, | |
| "think_blocks": "stripped (no-think variant)", | |
| "final_loss": 1.0404357817745948, | |
| "runtime_seconds": 2157.1858 | |
| } |