Update README.md
Browse files
README.md
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
language:
|
| 4 |
-
- en
|
| 5 |
-
---
|
| 6 |
## Ling-2.6-flash: Faster Responses, Stronger Execution, Higher Token Efficiency
|
| 7 |
### Introduction
|
| 8 |
Today, we announce the official open-source release of **Ling-2.6-flash**, an **instruct model** with **104B total parameters** and **7.4B active parameters**.
|
|
@@ -69,7 +69,8 @@ uv venv ~/my_ling_env
|
|
| 69 |
|
| 70 |
source ~/my_ling_env/bin/activate
|
| 71 |
|
| 72 |
-
uv pip
|
|
|
|
| 73 |
```
|
| 74 |
|
| 75 |
##### Run Inference
|
|
@@ -169,4 +170,4 @@ Ling-2.6-flash has already made meaningful progress in our pursuit of an extreme
|
|
| 169 |
|
| 170 |
At the same time, we are fully aware that pushing intelligence efficiency to the limit comes with tradeoffs. In some highly complex scenarios, the model can still exhibit **tool hallucinations** due to limited reasoning depth. In addition, there is still room for improvement in areas such as **natural bilingual switching between Chinese and English** and **compliance with highly complex instructions**.
|
| 171 |
|
| 172 |
-
Looking ahead, we will continue exploring the frontier of intelligence efficiency. While preserving the model’s high-efficiency inference characteristics, we aim to further improve the balance between **output quality** and **token efficiency**, and to continuously strengthen the model’s **stability, usability, and interaction experience across a wider range of real-world scenarios**.
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
---
|
| 6 |
## Ling-2.6-flash: Faster Responses, Stronger Execution, Higher Token Efficiency
|
| 7 |
### Introduction
|
| 8 |
Today, we announce the official open-source release of **Ling-2.6-flash**, an **instruct model** with **104B total parameters** and **7.4B active parameters**.
|
|
|
|
| 69 |
|
| 70 |
source ~/my_ling_env/bin/activate
|
| 71 |
|
| 72 |
+
# uv pip "sglang-kernel>=0.4.1"
|
| 73 |
+
uv pip install "sglang>=0.5.10.post1"
|
| 74 |
```
|
| 75 |
|
| 76 |
##### Run Inference
|
|
|
|
| 170 |
|
| 171 |
At the same time, we are fully aware that pushing intelligence efficiency to the limit comes with tradeoffs. In some highly complex scenarios, the model can still exhibit **tool hallucinations** due to limited reasoning depth. In addition, there is still room for improvement in areas such as **natural bilingual switching between Chinese and English** and **compliance with highly complex instructions**.
|
| 172 |
|
| 173 |
+
Looking ahead, we will continue exploring the frontier of intelligence efficiency. While preserving the model’s high-efficiency inference characteristics, we aim to further improve the balance between **output quality** and **token efficiency**, and to continuously strengthen the model’s **stability, usability, and interaction experience across a wider range of real-world scenarios**.
|