Update README.md
Browse files
README.md
CHANGED
|
@@ -37,7 +37,7 @@ For more details on how we trained this model, including on data automixing and
|
|
| 37 |
- **KV cache in FP8**: All quantization formats use a KV cache quantized to FP8, reducing memory per token
|
| 38 |
- **Native reasoning support**: Interleaved thinking enabled by default
|
| 39 |
- **Local-ready**: At 33B total parameters and 3B activated, Laguna XS.2 is compact enough to run on a Mac with 36 GB of RAM. [Available on Ollama](https://ollama.com/library/laguna-xs.2)
|
| 40 |
-
- **Apache 2.0 license**: Use and modify freely for
|
| 41 |
|
| 42 |
---
|
| 43 |
|
|
@@ -120,7 +120,7 @@ Submit feedback with `/feedback` and read the [full documentation on GitHub](htt
|
|
| 120 |
|
| 121 |
*By downloading and using pool, you agree to the Poolside [End User License Agreement (EULA)](https://poolside.ai/legal/eula).*
|
| 122 |
|
| 123 |
-
##
|
| 124 |
|
| 125 |
[vLLM, Transformers v5, TRT-LLM, SGLang, ...]
|
| 126 |
|
|
@@ -128,15 +128,15 @@ Thanks to support from Ollama and the mlx-lm team...
|
|
| 128 |
|
| 129 |
[Device frameworks: Ollama, mlx-lm, ...]
|
| 130 |
|
| 131 |
-
###
|
| 132 |
|
| 133 |
[...]
|
| 134 |
|
| 135 |
-
###
|
| 136 |
|
| 137 |
[...]
|
| 138 |
|
| 139 |
-
###
|
| 140 |
|
| 141 |
## Controlling reasoning
|
| 142 |
|
|
|
|
| 37 |
- **KV cache in FP8**: All quantization formats use a KV cache quantized to FP8, reducing memory per token
|
| 38 |
- **Native reasoning support**: Interleaved thinking enabled by default
|
| 39 |
- **Local-ready**: At 33B total parameters and 3B activated, Laguna XS.2 is compact enough to run on a Mac with 36 GB of RAM. [Available on Ollama](https://ollama.com/library/laguna-xs.2)
|
| 40 |
+
- **Apache 2.0 license**: Use and modify freely for commercial and non-commercial purposes
|
| 41 |
|
| 42 |
---
|
| 43 |
|
|
|
|
| 120 |
|
| 121 |
*By downloading and using pool, you agree to the Poolside [End User License Agreement (EULA)](https://poolside.ai/legal/eula).*
|
| 122 |
|
| 123 |
+
## Local deployment
|
| 124 |
|
| 125 |
[vLLM, Transformers v5, TRT-LLM, SGLang, ...]
|
| 126 |
|
|
|
|
| 128 |
|
| 129 |
[Device frameworks: Ollama, mlx-lm, ...]
|
| 130 |
|
| 131 |
+
### vLLM
|
| 132 |
|
| 133 |
[...]
|
| 134 |
|
| 135 |
+
### Transformers
|
| 136 |
|
| 137 |
[...]
|
| 138 |
|
| 139 |
+
### [Other frameworks]
|
| 140 |
|
| 141 |
## Controlling reasoning
|
| 142 |
|