Update README.md
Browse files
README.md
CHANGED
|
@@ -58,9 +58,12 @@ Check out our [llama.cpp documentation](https://qwen.readthedocs.io/en/latest/ru
|
|
| 58 |
We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp.
|
| 59 |
In the following demonstration, we assume that you are running commands under the repository `llama.cpp`.
|
| 60 |
|
|
|
|
|
|
|
|
|
|
| 61 |
|
| 62 |
```shell
|
| 63 |
-
./llama-cli -
|
| 64 |
```
|
| 65 |
|
| 66 |
## Processing Long Texts
|
|
|
|
| 58 |
We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp.
|
| 59 |
In the following demonstration, we assume that you are running commands under the repository `llama.cpp`.
|
| 60 |
|
| 61 |
+
```shell
|
| 62 |
+
huggingface-cli download Qwen/Qwen3-Coder-Next-GGUF --include "Qwen3-Coder-Next-Q5_K_M/*"
|
| 63 |
+
```
|
| 64 |
|
| 65 |
```shell
|
| 66 |
+
./llama-cli -m ./Qwen3-Coder-Next-Q5_K_M/Qwen3-Coder-Next-00001-of-00004.gguf --jinja -ngl 99 -fa on -sm row --temp 1.0 --top-k 40 --top-p 0.95 --min-p 0 -c 40960 -n 32768 --no-context-shift
|
| 67 |
```
|
| 68 |
|
| 69 |
## Processing Long Texts
|