varunrandery commited on
Commit
a4fde8e
·
verified ·
1 Parent(s): 0fb0c2d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -37,7 +37,7 @@ For more details on how we trained this model, including on data automixing and
37
  - **KV cache in FP8**: All quantization formats use a KV cache quantized to FP8, reducing memory per token
38
  - **Native reasoning support**: Interleaved thinking enabled by default
39
  - **Local-ready**: At 33B total parameters and 3B activated, Laguna XS.2 is compact enough to run on a Mac with 36 GB of RAM. [Available on Ollama](https://ollama.com/library/laguna-xs.2)
40
- - **Apache 2.0 license**: Use and modify freely for commerical and non-commercial purposes
41
 
42
  ---
43
 
@@ -120,7 +120,7 @@ Submit feedback with `/feedback` and read the [full documentation on GitHub](htt
120
 
121
  *By downloading and using pool, you agree to the Poolside [End User License Agreement (EULA)](https://poolside.ai/legal/eula).*
122
 
123
- ### Local deployment
124
 
125
  [vLLM, Transformers v5, TRT-LLM, SGLang, ...]
126
 
@@ -128,15 +128,15 @@ Thanks to support from Ollama and the mlx-lm team...
128
 
129
  [Device frameworks: Ollama, mlx-lm, ...]
130
 
131
- #### vLLM
132
 
133
  [...]
134
 
135
- #### Transformers
136
 
137
  [...]
138
 
139
- #### [Other frameworks]
140
 
141
  ## Controlling reasoning
142
 
 
37
  - **KV cache in FP8**: All quantization formats use a KV cache quantized to FP8, reducing memory per token
38
  - **Native reasoning support**: Interleaved thinking enabled by default
39
  - **Local-ready**: At 33B total parameters and 3B activated, Laguna XS.2 is compact enough to run on a Mac with 36 GB of RAM. [Available on Ollama](https://ollama.com/library/laguna-xs.2)
40
+ - **Apache 2.0 license**: Use and modify freely for commercial and non-commercial purposes
41
 
42
  ---
43
 
 
120
 
121
  *By downloading and using pool, you agree to the Poolside [End User License Agreement (EULA)](https://poolside.ai/legal/eula).*
122
 
123
+ ## Local deployment
124
 
125
  [vLLM, Transformers v5, TRT-LLM, SGLang, ...]
126
 
 
128
 
129
  [Device frameworks: Ollama, mlx-lm, ...]
130
 
131
+ ### vLLM
132
 
133
  [...]
134
 
135
+ ### Transformers
136
 
137
  [...]
138
 
139
+ ### [Other frameworks]
140
 
141
  ## Controlling reasoning
142