| { |
| "base_model_id": "HuggingFaceH4/zephyr-7b-beta", |
| "adapters": { |
| "adapter_1": "lamm-mit/x-lora/adapter_1", |
| "adapter_2": "lamm-mit/x-lora/adapter_2/", |
| "adapter_3": "lamm-mit/x-lora/adapter_3/", |
| "adapter_4": "lamm-mit/x-lora/adapter_4/", |
| "adapter_5": "lamm-mit/x-lora/adapter_5/", |
| "adapter_6": "lamm-mit/x-lora/adapter_6/", |
| "adapter_7": "lamm-mit/x-lora/adapter_7/", |
| "adapter_8": "lamm-mit/x-lora/adapter_8/", |
| "adapter_9": "lamm-mit/x-lora/adapter_9/" |
| }, |
| "hidden_size": 4096, |
| "enable_softmax": true, |
| "enable_softmax_topk": false, |
| "layerwise_scalings": true, |
| "xlora_depth": 1, |
| "xlora_size": 2048, |
| "enable_relu_and_dropout": true, |
| "use_bias": true, |
| "xlora_dropout_p": 0.2, |
| "use_trainable_adapters": false, |
| "softmax_temperature": 1, |
| "top_k_lora": null, |
| "scaling_pass_value": 0 |
| } |