caizhi1 commited on
Commit
5e2e8d3
·
verified ·
1 Parent(s): 85b9552

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -7
README.md CHANGED
@@ -1,8 +1,8 @@
1
- ---
2
- license: mit
3
- language:
4
- - en
5
- ---
6
  ## Ling-2.6-flash: Faster Responses, Stronger Execution, Higher Token Efficiency
7
  ### Introduction
8
  Today, we announce the official open-source release of **Ling-2.6-flash**, an **instruct model** with **104B total parameters** and **7.4B active parameters**.
@@ -69,7 +69,8 @@ uv venv ~/my_ling_env
69
 
70
  source ~/my_ling_env/bin/activate
71
 
72
- uv pip install sglang
 
73
  ```
74
 
75
  ##### Run Inference
@@ -169,4 +170,4 @@ Ling-2.6-flash has already made meaningful progress in our pursuit of an extreme
169
 
170
  At the same time, we are fully aware that pushing intelligence efficiency to the limit comes with tradeoffs. In some highly complex scenarios, the model can still exhibit **tool hallucinations** due to limited reasoning depth. In addition, there is still room for improvement in areas such as **natural bilingual switching between Chinese and English** and **compliance with highly complex instructions**.
171
 
172
- Looking ahead, we will continue exploring the frontier of intelligence efficiency. While preserving the model’s high-efficiency inference characteristics, we aim to further improve the balance between **output quality** and **token efficiency**, and to continuously strengthen the model’s **stability, usability, and interaction experience across a wider range of real-world scenarios**.
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ ---
6
  ## Ling-2.6-flash: Faster Responses, Stronger Execution, Higher Token Efficiency
7
  ### Introduction
8
  Today, we announce the official open-source release of **Ling-2.6-flash**, an **instruct model** with **104B total parameters** and **7.4B active parameters**.
 
69
 
70
  source ~/my_ling_env/bin/activate
71
 
72
+ # uv pip "sglang-kernel>=0.4.1"
73
+ uv pip install "sglang>=0.5.10.post1"
74
  ```
75
 
76
  ##### Run Inference
 
170
 
171
  At the same time, we are fully aware that pushing intelligence efficiency to the limit comes with tradeoffs. In some highly complex scenarios, the model can still exhibit **tool hallucinations** due to limited reasoning depth. In addition, there is still room for improvement in areas such as **natural bilingual switching between Chinese and English** and **compliance with highly complex instructions**.
172
 
173
+ Looking ahead, we will continue exploring the frontier of intelligence efficiency. While preserving the model’s high-efficiency inference characteristics, we aim to further improve the balance between **output quality** and **token efficiency**, and to continuously strengthen the model’s **stability, usability, and interaction experience across a wider range of real-world scenarios**.