DedeProGames commited on
Commit
f4a619f
·
verified ·
1 Parent(s): f52be54

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +222 -0
README.md ADDED
@@ -0,0 +1,222 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - Qwen/Qwen3.6-27B
5
+ ---
6
+
7
+ <p align="center">
8
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/685ea8ff7b4139b6845ce395/_66bkNH630dGeIt2Uuctd.png" alt="logo" width="500">
9
+ </p>
10
+ <div align="center">
11
+ <a href="https://huggingface.co/OrionLLM/GRM-2.6-Plus/" style="text-decoration: none;">
12
+ <img src="https://img.shields.io/badge/🤗-HuggingFace-FC926C?style=for-the-badge" alt="HuggingFace">
13
+ </a>
14
+ <a href="https://huggingface.co/collections/OrionLLM/grm-26" style="text-decoration: none;">
15
+ <img src="https://img.shields.io/badge/📚-Collection-3B82F6?style=for-the-badge" alt="Collection">
16
+ </a>
17
+ <a href="https://www.apache.org/licenses/LICENSE-2.0" style="text-decoration: none;">
18
+ <img src="https://img.shields.io/badge/📜-License-E343BD?style=for-the-badge" alt="License">
19
+ </a>
20
+ </div>
21
+
22
+ ## 1. Introduction
23
+
24
+ GRM-2.6-Plus is a **27B-parameter reasoning model** built for **general-purpose AI** and optimized for **difficult, high-complexity tasks**. It is designed to deliver stronger performance for its size while remaining practical, efficient, and accessible for advanced local and research-oriented use.
25
+
26
+ The model focuses on **structured reasoning**, helping it produce more accurate, coherent, and reliable responses across demanding problems. GRM-2.6-Plus brings **elite-level reasoning** to complex workloads, making it suitable for users who need a capable model for advanced problem-solving, coding, agents, and everyday intelligence.
27
+
28
+ ## 2. Key Capabilities
29
+
30
+ - **Elite-Level Reasoning for Hard Tasks:** GRM-2.6-Plus is optimized to handle difficult reasoning workloads with clarity, consistency, and strong step-by-step problem-solving ability.
31
+ - **High Performance for Its Size:** With **27B parameters**, the model is designed to deliver excellent capability relative to its scale, balancing strong intelligence with practical deployment.
32
+ - **Advanced Coding and Agentic Use:** GRM-2.6-Plus is well suited for code generation, structured problem-solving, tool-style workflows, and local agentic applications.
33
+ - **Optimized for Practical Deployment:** The model aims to remain efficient and usable across capable consumer and workstation hardware while offering strong performance for advanced tasks.
34
+
35
+ ## 3. Performance
36
+
37
+ GRM-2.6-Plus is designed to be a highly capable **27B local AI model** for complex reasoning, coding, everyday chat, and agentic workflows. It focuses on delivering **better performance for its size**, making it a strong option for users who want powerful reasoning without relying only on massive-scale models.
38
+
39
+ Its core strength is **practical intelligence**: elite-level reasoning, strong task understanding, stable responses, and the ability to handle difficult problems across multiple domains.
40
+
41
+ <table>
42
+ <tr>
43
+ <th style="background: rgba(128,128,128,0.1); text-align: center;"> </th>
44
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">GRM-2.6-Plus</th>
45
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">Qwen3.6-27B</th>
46
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">google/gemma-4-31B-it</th>
47
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">GPT-5.4-Mini</th>
48
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">Claude-4.5-Haiku</th>
49
+ </tr>
50
+ <tr>
51
+ <td align="center" colspan="6" style="background: linear-gradient(90deg, rgba(124,58,237,0.45) 0%, rgba(99,102,241,0.42) 50%, rgba(59,130,246,0.45) 100%); font-weight: bold; height:32px; padding-top:2px; padding-bottom:2px;"><i>Knowledge &amp; STEM</i></td>
52
+ </tr>
53
+ <tr>
54
+ <td align="center">MMLU-Pro</td>
55
+ <td align="center"><b>86.8</b></td>
56
+ <td align="center">86.2</td>
57
+ <td align="center">85.2</td>
58
+ <td align="center">--</td>
59
+ <td align="center">80.0</td>
60
+ </tr>
61
+ <tr>
62
+ <td align="center">MMLU-Redux</td>
63
+ <td align="center"><b>94.2</b></td>
64
+ <td align="center">93.5</td>
65
+ <td align="center">93.7</td>
66
+ <td align="center">--</td>
67
+ <td align="center">--</td>
68
+ </tr>
69
+ <tr>
70
+ <td align="center">C-Eval</td>
71
+ <td align="center"><b>92.0</b></td>
72
+ <td align="center">91.4</td>
73
+ <td align="center">82.6</td>
74
+ <td align="center">--</td>
75
+ <td align="center">--</td>
76
+ </tr>
77
+ <tr>
78
+ <td align="center">GPQA Diamond</td>
79
+ <td align="center"><b>88.3</b></td>
80
+ <td align="center">87.8</td>
81
+ <td align="center">84.3</td>
82
+ <td align="center">88.0</td>
83
+ <td align="center">73.0</td>
84
+ </tr>
85
+ <tr>
86
+ <td align="center">SuperGPQA</td>
87
+ <td align="center"><b>66.4</b></td>
88
+ <td align="center">66.0</td>
89
+ <td align="center">65.7</td>
90
+ <td align="center">--</td>
91
+ <td align="center">--</td>
92
+ </tr>
93
+ <tr>
94
+ <td align="center" colspan="6" style="background: linear-gradient(90deg, rgba(124,58,237,0.45) 0%, rgba(99,102,241,0.42) 50%, rgba(59,130,246,0.45) 100%); font-weight: bold; height:32px; padding-top:2px; padding-bottom:2px;"><i>Reasoning &amp; Coding</i></td>
95
+ </tr>
96
+ <tr>
97
+ <td align="center">LiveCodeBench v6</td>
98
+ <td align="center"><b>84.8</b></td>
99
+ <td align="center">83.9</td>
100
+ <td align="center">80.0</td>
101
+ <td align="center">--</td>
102
+ <td align="center">51.1</td>
103
+ </tr>
104
+ <tr>
105
+ <td align="center">HMMT Feb 26</td>
106
+ <td align="center"><b>84.8</b></td>
107
+ <td align="center">84.3</td>
108
+ <td align="center">77.2</td>
109
+ <td align="center">--</td>
110
+ <td align="center">--</td>
111
+ </tr>
112
+ <tr>
113
+ <td align="center">AIME26</td>
114
+ <td align="center"><b>95.1</b></td>
115
+ <td align="center">94.1</td>
116
+ <td align="center">89.2</td>
117
+ <td align="center">--</td>
118
+ <td align="center">--</td>
119
+ </tr>
120
+ <tr>
121
+ <td align="center" colspan="6" style="background: linear-gradient(90deg, rgba(124,58,237,0.45) 0%, rgba(99,102,241,0.42) 50%, rgba(59,130,246,0.45) 100%); font-weight: bold; height:32px; padding-top:2px; padding-bottom:2px;"><i>General Agent</i></td>
122
+ </tr>
123
+ <tr>
124
+ <td align="center">SWE-bench Verified</td>
125
+ <td align="center"><b>78.7</b></td>
126
+ <td align="center">77.2</td>
127
+ <td align="center">52.0</td>
128
+ <td align="center">--</td>
129
+ <td align="center">73.3</td>
130
+ </tr>
131
+ <tr>
132
+ <td align="center">SWE-bench Pro</td>
133
+ <td align="center"><b>54.0</b></td>
134
+ <td align="center">53.5</td>
135
+ <td align="center">35.7</td>
136
+ <td align="center">54.4</td>
137
+ <td align="center">--</td>
138
+ </tr>
139
+ <tr>
140
+ <td align="center">Terminal-Bench 2.0</td>
141
+ <td align="center"><b>59.8</b></td>
142
+ <td align="center">59.3</td>
143
+ <td align="center">42.9</td>
144
+ <td align="center">60.0</td>
145
+ <td align="center">41.0</td>
146
+ </tr>
147
+ </table>
148
+
149
+ ## 4. Family
150
+ The GRM-2.6 family is available in various sizes to suit every case.
151
+
152
+ <table>
153
+ <tr>
154
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">Model</th>
155
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">Size</th>
156
+ <th style="background: rgba(128,128,128,0.1); text-align: center;">Domain</th>
157
+ </tr>
158
+ <tr>
159
+ <td align="center">GRM-2.6-Plus</td>
160
+ <td align="center">27B</td>
161
+ <td align="center">Powerful model for extremely difficult tasks</td>
162
+ </tr>
163
+ <tr>
164
+ <td align="center">GRM-2.6</td>
165
+ <td align="center">9B</td>
166
+ <td align="center">Powerful on-device deployment for difficult tasks</td>
167
+ </tr>
168
+ <tr>
169
+ <td align="center">GRM-2.6-Air</td>
170
+ <td align="center">2B</td>
171
+ <td align="center">Any-device deployment for everyday chat</td>
172
+ </tr>
173
+ </table>
174
+
175
+ ## 5. Architecture
176
+ GRM-2.6 is built on the Qwen3.6 architecture and is optimized for complex tasks, agent environments, and everyday chat.
177
+
178
+ GRM-2.6 applies the same principle to a stronger, larger foundation, resulting in a model that punches above its weight class on structured reasoning tasks while remaining deployable on consumer hardware.
179
+
180
+ ## 6. Quick start
181
+
182
+ Before starting, make sure it is installed and the API key and the API base URL is configured, e.g.:
183
+ ```shell
184
+ pip install -U openai
185
+
186
+ # Set the following accordingly
187
+ export OPENAI_BASE_URL="http://localhost:8000/v1"
188
+ export OPENAI_API_KEY="EMPTY"
189
+ ```
190
+
191
+ #### Text-Only Input
192
+
193
+ ```python
194
+ from openai import OpenAI
195
+ # Configured by environment variables
196
+ client = OpenAI()
197
+
198
+ messages = [
199
+ {"role": "user", "content": "Create an calculator in a single HTML file backwards"},
200
+ ]
201
+
202
+ chat_response = client.chat.completions.create(
203
+ model="OrionLLM/GRM-2.5-Plus",
204
+ messages=messages,
205
+ max_tokens=81920,
206
+ temperature=1.0,
207
+ top_p=0.95,
208
+ presence_penalty=0.0,
209
+ extra_body={
210
+ "top_k": 20,
211
+ },
212
+ )
213
+ print("Chat response:", chat_response)
214
+ ```
215
+
216
+ ---
217
+
218
+ <div align="center">
219
+
220
+ **GRM-2.6-Plus** is developed by **[OrionLLM](https://huggingface.co/OrionLLM)** and released under the Apache 2.0 License.
221
+
222
+ </div>