| { |
| "adapters": { |
| "0": "msfm/llm-jp-3-1.8b_adapter-1", |
| "1": "msfm/llm-jp-3-1.8b_adapter-2" |
| }, |
| "auto_mapping": null, |
| "base_model_name_or_path": "llm-jp/llm-jp-3-1.8b", |
| "enable_softmax": true, |
| "enable_softmax_topk": false, |
| "global_scaling_weight": 1.0, |
| "hidden_size": 2048, |
| "inference_mode": true, |
| "layerwise_scalings": false, |
| "peft_type": "XLORA", |
| "revision": null, |
| "scaling_pass_value": 0.0, |
| "softmax_temperature": 1.0, |
| "task_type": "CAUSAL_LM", |
| "top_k_lora": null, |
| "use_trainable_adapters": false, |
| "xlora_depth": 4, |
| "xlora_dropout_p": 0.2, |
| "xlora_size": 2048 |
| } |