| |
| id: smollm2 |
| model: smollm2 |
| name: smollm2 |
| version: 1 |
| |
|
|
| |
| |
| stop: |
| - <|im_end|> |
| |
|
|
| |
| stream: true |
| top_p: 0.9 |
| temperature: 0.7 |
| frequency_penalty: 0 |
| presence_penalty: 0 |
| max_tokens: 4096 |
| seed: -1 |
| dynatemp_range: 0 |
| dynatemp_exponent: 1 |
| top_k: 40 |
| min_p: 0.05 |
| tfs_z: 1 |
| typ_p: 1 |
| repeat_last_n: 64 |
| repeat_penalty: 1 |
| mirostat: false |
| mirostat_tau: 5 |
| mirostat_eta: 0.100000001 |
| penalize_nl: false |
| ignore_eos: false |
| n_probs: 0 |
| min_keep: 0 |
| |
| |
|
|
| |
| |
| engine: llama-cpp |
| ctx_len: 4096 |
| ngl: 34 |
| prompt_template: "<|im_start|>system |
| {system_message}<|im_end|> |
| <|im_start|>user |
| {prompt}<|im_end|> |
| <|im_start|>assistant" |
| |
| |
|
|