Dataset Viewer
Auto-converted to Parquet Duplicate
prompt_id
large_stringclasses
499 values
source
large_stringclasses
5 values
text
large_stringclasses
497 values
text_len_chars
int64
2
92.3k
text_len_words
int64
1
10k
language
large_stringclasses
2 values
domain
large_stringclasses
3 values
collected_at
large_stringdate
2026-05-11 00:32:32
2026-05-11 00:32:37
provider
large_stringclasses
3 values
format
large_stringclasses
5 values
model
large_stringclasses
3 values
offline_count
int64
1
36.9k
tokenizer_version
large_stringclasses
3 values
offline_ts
large_stringdate
2026-05-11 01:06:36
2026-05-11 01:12:25
empirical_count
int64
1
45.5k
is_oracle
bool
1 class
empirical_source
large_stringclasses
3 values
endpoint
large_stringclasses
3 values
empirical_ts
large_stringdate
2026-05-11 01:13:54
2026-05-11 01:50:38
delta
int64
-126
18.5k
delta_pct
float64
-100
92.3
abs_delta
int64
0
18.5k
direction
large_stringclasses
3 values
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
anthropic
plain
claude-opus-4-7
83
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
146
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:25Z
63
43.150685
63
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
openai
plain
gpt-4o
85
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
85
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
mistral
plain
mistral-large-latest
107
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
106
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.943396
1
overestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
anthropic
markdown
claude-opus-4-7
86
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
152
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:25Z
66
43.421053
66
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
openai
markdown
gpt-4o
88
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
88
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
mistral
markdown
mistral-large-latest
112
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
111
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.900901
1
overestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
anthropic
json
claude-opus-4-7
85
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
190
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:25Z
105
55.263158
105
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
openai
json
gpt-4o
87
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
110
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
23
20.909091
23
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
mistral
json
mistral-large-latest
109
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
136
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
27
19.852941
27
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
anthropic
xml
claude-opus-4-7
156
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
258
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:25Z
102
39.534884
102
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
openai
xml
gpt-4o
152
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
152
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
mistral
xml
mistral-large-latest
178
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
177
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.564972
1
overestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
anthropic
yaml
claude-opus-4-7
85
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
156
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:25Z
71
45.512821
71
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
openai
yaml
gpt-4o
87
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
91
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
4
4.395604
4
underestimate
2dbc2134-c30f-446e-8026-695ff8cda88b
humaneval
def correct_bracketing(brackets: str): """ brackets is a string of "(" and ")". return True if every opening bracket has a corresponding closing bracket. >>> correct_bracketing("(") False >>> correct_bracketing("()") True >>> correct_bracketing("(()())") True >>> correct_bracketing(...
345
36
code
code
2026-05-11T00:32:32Z
mistral
yaml
mistral-large-latest
109
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
114
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
5
4.385965
5
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
anthropic
plain
claude-opus-4-7
96
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
147
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
51
34.693878
51
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
openai
plain
gpt-4o
95
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
95
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
mistral
plain
mistral-large-latest
132
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
131
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.763359
1
overestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
anthropic
markdown
claude-opus-4-7
99
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
153
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
54
35.294118
54
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
openai
markdown
gpt-4o
98
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
98
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
mistral
markdown
mistral-large-latest
137
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
136
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.735294
1
overestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
anthropic
json
claude-opus-4-7
98
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
176
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
78
44.318182
78
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
openai
json
gpt-4o
97
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
119
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
22
18.487395
22
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
mistral
json
mistral-large-latest
134
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
147
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
13
8.843537
13
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
anthropic
xml
claude-opus-4-7
126
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
201
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
75
37.313433
75
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
openai
xml
gpt-4o
126
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
126
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
mistral
xml
mistral-large-latest
165
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
164
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.609756
1
overestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
anthropic
yaml
claude-opus-4-7
98
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
157
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
59
37.579618
59
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
openai
yaml
gpt-4o
97
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
101
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
4
3.960396
4
underestimate
deda4e16-1b3d-4d5c-a9a1-fa6f81f76d1c
humaneval
def unique_digits(x): """Given a list of positive integers x. return a sorted list of all elements that hasn't any even digit. Note: Returned list should be sorted in increasing order. For example: >>> unique_digits([15, 33, 1422, 1]) [1, 15, 33] >>> unique_digits([152, 323, 1422, 10]...
336
47
code
code
2026-05-11T00:32:32Z
mistral
yaml
mistral-large-latest
134
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
139
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
5
3.597122
5
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
anthropic
plain
claude-opus-4-7
270
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
370
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
100
27.027027
100
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
openai
plain
gpt-4o
270
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
270
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
mistral
plain
mistral-large-latest
307
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
306
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.326797
1
overestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
anthropic
markdown
claude-opus-4-7
273
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
376
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
103
27.393617
103
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
openai
markdown
gpt-4o
273
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
273
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
mistral
markdown
mistral-large-latest
312
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
311
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.321543
1
overestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
anthropic
json
claude-opus-4-7
272
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
442
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
170
38.461538
170
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
openai
json
gpt-4o
272
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
308
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
36
11.688312
36
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
mistral
json
mistral-large-latest
309
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
331
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
22
6.646526
22
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
anthropic
xml
claude-opus-4-7
368
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
551
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
183
33.212341
183
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
openai
xml
gpt-4o
368
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
368
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
mistral
xml
mistral-large-latest
442
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
441
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.226757
1
overestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
anthropic
yaml
claude-opus-4-7
272
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
380
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:26Z
108
28.421053
108
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
openai
yaml
gpt-4o
272
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
276
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
4
1.449275
4
underestimate
a39231a7-d777-4477-8c66-e0a8a013ac6e
humaneval
def by_length(arr): """ Given an array of integers, sort the integers that are between 1 and 9 inclusive, reverse the resulting array, and then replace each digit by its corresponding name from "One", "Two", "Three", "Four", "Five", "Six", "Seven", "Eight", "Nine". For example: arr = [2, 1, 1...
807
133
code
code
2026-05-11T00:32:32Z
mistral
yaml
mistral-large-latest
309
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
314
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
5
1.592357
5
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
anthropic
plain
claude-opus-4-7
102
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
134
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
32
23.880597
32
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
openai
plain
gpt-4o
103
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
103
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
mistral
plain
mistral-large-latest
122
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
121
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.826446
1
overestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
anthropic
markdown
claude-opus-4-7
105
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
140
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
35
25
35
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
openai
markdown
gpt-4o
106
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
106
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
mistral
markdown
mistral-large-latest
127
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
126
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.793651
1
overestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
anthropic
json
claude-opus-4-7
104
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
158
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
54
34.177215
54
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
openai
json
gpt-4o
105
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
120
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
15
12.5
15
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
mistral
json
mistral-large-latest
124
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
134
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
10
7.462687
10
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
anthropic
xml
claude-opus-4-7
131
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
184
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
53
28.804348
53
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
openai
xml
gpt-4o
132
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
132
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
mistral
xml
mistral-large-latest
153
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
152
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.657895
1
overestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
anthropic
yaml
claude-opus-4-7
104
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
142
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
38
26.760563
38
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
openai
yaml
gpt-4o
105
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
108
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
3
2.777778
3
underestimate
2720797d-32eb-4689-9be5-78c781f631d4
humaneval
def incr_list(l: list): """Return list with elements incremented by 1. >>> incr_list([1, 2, 3]) [2, 3, 4] >>> incr_list([5, 3, 5, 2, 3, 3, 9, 0, 123]) [6, 4, 6, 3, 4, 4, 10, 1, 124] """
209
37
code
code
2026-05-11T00:32:32Z
mistral
yaml
mistral-large-latest
124
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
128
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
4
3.125
4
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
anthropic
plain
claude-opus-4-7
109
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
155
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
46
29.677419
46
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
openai
plain
gpt-4o
109
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
109
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
mistral
plain
mistral-large-latest
137
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
136
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.735294
1
overestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
anthropic
markdown
claude-opus-4-7
112
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
161
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
49
30.434783
49
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
openai
markdown
gpt-4o
112
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
112
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
mistral
markdown
mistral-large-latest
142
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
141
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.70922
1
overestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
anthropic
json
claude-opus-4-7
111
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
184
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
73
39.673913
73
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
openai
json
gpt-4o
111
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
128
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
17
13.28125
17
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
mistral
json
mistral-large-latest
139
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
151
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
12
7.94702
12
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
anthropic
xml
claude-opus-4-7
138
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
205
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
67
32.682927
67
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
openai
xml
gpt-4o
138
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
138
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
mistral
xml
mistral-large-latest
168
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
167
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.598802
1
overestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
anthropic
yaml
claude-opus-4-7
111
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
165
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:27Z
54
32.727273
54
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
openai
yaml
gpt-4o
111
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
115
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
4
3.478261
4
underestimate
8a14be62-295b-4715-8333-e8615fb8d16c
humaneval
def order_by_points(nums): """ Write a function which sorts the given list of integers in ascending order according to the sum of their digits. Note: if there are several items with similar sum of their digits, order them based on their index in original list. For example: >>> order_by_poin...
412
63
code
code
2026-05-11T00:32:32Z
mistral
yaml
mistral-large-latest
139
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
144
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
5
3.472222
5
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
anthropic
plain
claude-opus-4-7
229
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
330
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
101
30.606061
101
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
openai
plain
gpt-4o
228
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
228
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
mistral
plain
mistral-large-latest
261
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
260
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.384615
1
overestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
anthropic
markdown
claude-opus-4-7
232
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
336
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
104
30.952381
104
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
openai
markdown
gpt-4o
231
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
231
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
mistral
markdown
mistral-large-latest
266
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
265
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.377358
1
overestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
anthropic
json
claude-opus-4-7
231
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
362
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
131
36.187845
131
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
openai
json
gpt-4o
230
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
250
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
20
8
20
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
mistral
json
mistral-large-latest
263
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
274
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
11
4.014599
11
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
anthropic
xml
claude-opus-4-7
247
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
366
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
119
32.513661
119
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
openai
xml
gpt-4o
247
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
247
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
mistral
xml
mistral-large-latest
282
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
281
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.355872
1
overestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
anthropic
yaml
claude-opus-4-7
231
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
338
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
107
31.656805
107
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
openai
yaml
gpt-4o
230
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
233
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
3
1.287554
3
underestimate
87c5421e-ec24-43c5-8754-108ff4188f3f
humaneval
def tri(n): """Everyone knows Fibonacci sequence, it was studied deeply by mathematicians in the last couple centuries. However, what people don't know is Tribonacci sequence. Tribonacci sequence is defined by the recurrence: tri(1) = 3 tri(n) = 1 + n / 2, if n is even. tri(n) = tri(n - 1) + t...
670
126
code
code
2026-05-11T00:32:32Z
mistral
yaml
mistral-large-latest
263
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
267
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
4
1.498127
4
underestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
anthropic
plain
claude-opus-4-7
197
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
253
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
56
22.134387
56
underestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
openai
plain
gpt-4o
196
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
196
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
mistral
plain
mistral-large-latest
224
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
223
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.44843
1
overestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
anthropic
markdown
claude-opus-4-7
200
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
259
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
59
22.779923
59
underestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
openai
markdown
gpt-4o
199
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
199
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
0
0
0
exact
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
mistral
markdown
mistral-large-latest
229
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
228
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
-1
-0.438596
1
overestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
anthropic
json
claude-opus-4-7
199
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
287
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
88
30.662021
88
underestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
openai
json
gpt-4o
198
o200k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
220
true
tiktoken
tiktoken==0.12.0 encoding=o200k_base model=gpt-4o
2026-05-11T01:13:54Z
22
10
22
underestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
mistral
json
mistral-large-latest
226
mistral_v1_v3@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
238
true
sdk
mistral-common==1.11.2 tokenizer=SentencePieceTokenizer resolved=mistral-large-2411
2026-05-11T01:13:54Z
12
5.042017
12
underestimate
52fbe43b-9954-4eb4-8025-7ad1eb2263dd
humaneval
def sort_array(array): """ Given an array of non-negative integers, return a copy of the given array after sorting, you will sort the given array in ascending order if the sum( first index value, last index value) is odd, or sort it in descending order if the sum( first index value, last index value) is...
576
102
code
code
2026-05-11T00:32:32Z
anthropic
xml
claude-opus-4-7
223
cl100k_base@tokenometer-1.0.1+rates-2026-05-09
2026-05-11T01:06:36Z
305
true
api
anthropic.messages.count_tokens (sdk anthropic==0.100.0) model=claude-opus-4-7
2026-05-11T01:46:28Z
82
26.885246
82
underestimate
End of preview. Expand in Data Studio

Dataset Card for LLM Tokens Atlas

llm-tokens-atlas is an open, reproducible benchmark of LLM tokenization across 3 providers in v0.1.0 (Anthropic claude-opus-4-7, OpenAI gpt-4o, Mistral mistral-large-latest) across 5 prompt formats (Markdown, XML, JSON, YAML, Plain text), evaluated on 7,485 real-world prompt requests (n=2,495 per provider, 499 unique prompts × 5 formats × 3 providers). Google + Cohere parity sweeps are scheduled for v0.2.0. For each (prompt, provider, model, format) cell we record both an offline token count (from the provider's published or community-reverse-engineered tokenizer) and an empirical token count (from the provider's authoritative count-tokens API endpoint, where one exists).

The artifact is the calibration delta distribution between offline and empirical counts — per provider, per model, per format. It quantifies how wrong the cheap-to-run offline counter is, so anyone estimating cost or context-window budgets ahead of a real API call can correct for the bias rather than treat it as exact.

Coverage / What's in this release

Field Value
Version v0.1.0 (3-provider)
Release date 2026-05-11
Total rows 7,485
Unique prompts 499
Providers shipped 3 (Anthropic, OpenAI, Mistral)
Providers pending 2 (Google, Cohere) — schema-reserved, gated on v0.2.0 sweeps
Formats 5 (plain, markdown, json, xml, yaml)
Domains 3 (code, prose, chat)

Per-provider breakdown:

Provider Model n Empirical source
anthropic claude-opus-4-7 2,495 messages.countTokens (Anthropic API)
openai gpt-4o 2,495 tiktoken o200k_base (tiktoken-as-truth, local)
mistral mistral-large-latest 2,495 mistral-tokenizer-js (vendor OSS tokenizer)

Per-format breakdown (rows are split evenly across formats; each format gets 1,497 rows total = 499 prompts × 3 providers):

Format n total n per provider
plain 1,497 499
markdown 1,497 499
json 1,497 499
xml 1,497 499
yaml 1,497 499

Pending for v0.2.0:

  • Google gemini-2.5-pro via model.countTokens — gated on GOOGLE_API_KEY sweep.
  • Cohere command-r via POST /v1/tokenize — gated on COHERE_API_KEY sweep.

Both providers are reserved in the schema and validated end-to-end; the only missing piece is the empirical sweep. Rows for these providers will land in v0.2.0 without a major-version schema bump.

Headline empirical finding (v0.1.0):

  • Anthropic claude-opus-4-7: cl100k_base offline tokenizer underestimates empirical counts by 41.3% median (p25=36.8%, p75=46.7%, p95=58.6%; n=2,495). OLS calibration fit: slope = 1.611, intercept = 18.20, R² = 0.9956. 100% of rows underestimate (no exact or overestimate cases).
  • OpenAI gpt-4o: offline o200k_base is treated as oracle (tiktoken-as-truth); median delta = 0.0%, mean = 3.01%; calibration fit slope = 1.024, R² = 0.9986; 57.1% exact, 42.4% underestimate, 0.5% overestimate.
  • Mistral mistral-large-latest: median delta = −0.06% (mistral-tokenizer-js slightly overestimates by construction), mean = 1.85%; slope = 1.016, R² = 0.9993; 42.1% underestimate, 57.8% overestimate, ~0% exact.

Per-format × per-provider median deltas and full statistics live in analysis/results.json (the single source of truth for v0.1.0 numbers).

Dataset Summary

LLM Tokens Atlas measures the gap between what an offline tokenizer thinks a prompt costs and what a provider's API actually reports. Each row pairs a single prompt rendered in one of five surface formats with one provider model, records both counts, and computes the absolute and relative delta. Aggregated across providers and formats, the result is a calibrated joint distribution of offline-vs-empirical bias that downstream tools (cost estimators, routers, context budgeters, prompt optimizers) can plug in instead of guessing.

The dataset is intentionally narrow in task scope — it is a measurement artifact, not a training corpus. The prompts are drawn from already-public, redistribution-friendly corpora (LMSYS-Chat-1M sample, HumanEval, MT-Bench, GitHub READMEs, multilingual Wikipedia snippets). No model outputs are collected; only token counts.

Supported Tasks and Leaderboards

This dataset does not target a downstream supervised task. It is a measurement benchmark intended for:

  • Calibrating client-side / offline tokenizers against provider APIs.
  • Benchmarking cost-estimation libraries (e.g. tokencost, tokenometer).
  • Studying tokenizer drift over time as providers update their tokenizers.
  • Auditing context-window budgeting tools.
  • Grounding cost-aware routing systems in calibrated empirical token counts rather than offline proxies multiplied by published pricing.

task_categories: [other] reflects that this is a measurement / calibration artifact rather than a classical NLP supervised task.

Languages

The dataset's prompt text is overwhelmingly English (the LMSYS-Chat-1M sample and HumanEval contributions are English; multilingual Wikipedia snippets, when included, are tagged separately in the language field). For calibration purposes the underlying language matters less than the format and provider, but multilingual coverage is limited and explicitly so. See Limitations.

Dataset Structure

Data Files

A single Parquet file at data/processed/atlas.parquet is the canonical artifact. It is produced by llm_tokens_atlas/build_dataset.py, which inner-joins three intermediate JSONL streams — data/raw_prompts.jsonl, data/offline_counts.jsonl, and data/empirical_counts.jsonl — on the composite key (prompt_id, provider, format, model), then attaches prompt-level columns and computes calibration deltas. The full row-level schema for each intermediate stream is published as JSON Schema at data/schema.json; that file is the formal source of truth and should be preferred over this narrative description when they disagree.

Data Fields

The processed Parquet has the following columns:

Prompt-level (attached from raw_prompts.jsonl via prompt_id join):

Field Type Description
prompt_id string Stable identifier for the underlying prompt. Same prompt → same prompt_id across all (provider, model, format) rows; this is the join key downstream consumers should split on for held-out evaluation.
source string Origin corpus: lmsys-chat-1m, humaneval, mt-bench, github-readmes, wikipedia-multilingual, or similar (see data/provenance.md for the authoritative list).
text string The prompt text exactly as sent to the tokenizer / API. UTF-8.
text_len_chars int64 Length of text in Unicode codepoints (not bytes).
text_len_words int64 Whitespace-split word count of text. Approximate — for filtering / grouping, not as a tokenization signal.
language string ISO 639-1 / BCP-47 code: en, zh, es, fr, de, ja, multi, code, etc. code indicates source code; multi indicates mixed-language.
domain string High-level domain tag for stratified analysis: code, prose, chat, structured, multilingual, other.
collected_at string (ISO-8601) UTC timestamp at which the prompt was collected into the corpus.

Cell key:

Field Type Description
provider string One of: anthropic, openai, mistral (shipped in v0.1.0). Schema-reserved future values: google, cohere (rows land in v0.2.0).
format string Surface format the prompt is rendered in. One of: plain, markdown, xml, json, yaml.
model string Concrete model identifier evaluated. v0.1.0 ships three: claude-opus-4-7, gpt-4o, mistral-large-latest. Schema also reserves gemini-2.5-pro (Google) and command-r (Cohere) for v0.2.0.

Offline counts (from offline_counts.jsonl):

Field Type Description
offline_count int64 Token count from the offline tokenizer (tiktoken proxy, published BPE vocab, community-reverse-engineered tokenizer, or mistral-common).
tokenizer_version string Pinned tokenizer version identifier (e.g. tiktoken@cl100k_base, @tokenometer/core@1.0.0, mistral-common@1.7.0). Pinned for reproducibility in data/lockfile.json.
offline_ts string (ISO-8601) UTC timestamp at which the offline count was produced.

Empirical counts (from empirical_counts.jsonl):

Field Type Description
empirical_count int64 Token count from the provider's authoritative count-tokens endpoint (or tiktoken-as-truth for OpenAI, where it is treated as the oracle).
is_oracle bool True when this value is the ground-truth oracle for the provider (e.g. tiktoken for OpenAI, provider countTokens API for Anthropic / Google). False when it is empirical but not official (e.g. inferred from stream usage).
empirical_source string How the empirical count was obtained: api (HTTP call), tiktoken (local oracle), sdk (vendor SDK helper), or stream-usage (inferred from generation metadata).
endpoint string Concrete endpoint or library identifier (e.g. https://api.anthropic.com/v1/messages/count_tokens@2024-10-22, tiktoken.encoding_for_model(gpt-4o)).
empirical_ts string (ISO-8601) UTC timestamp at which the empirical count was produced.

Computed calibration columns (added at build time):

Field Type Description
delta int64 empirical_count − offline_count. Positive ⇒ offline underestimates the real count; negative ⇒ overestimates; zero ⇒ exact agreement.
abs_delta int64 abs(delta). Useful for symmetric error metrics.
delta_pct float64 delta / empirical_count × 100. The headline per-row calibration error, in percent. NaN where empirical_count == 0.
direction string Categorical mapping of sign(delta): underestimate, overestimate, or exact.

Data Splits

The dataset is published as a single train split. There is no held-out evaluation set — this is a calibration / measurement artifact, not a supervised task. Downstream users who need a held-out set should split by prompt_id so that the same prompt rendered in different formats does not leak across splits.

Dataset Creation

Curation Rationale

A handful of independent observations in 2026 — most prominently two blog posts documenting that Anthropic's updated tokenizer inflates token counts by ~40–47% on real prompts — surfaced the same underlying phenomenon: deployed offline tokenizers consistently disagree with provider APIs, and the disagreement is large enough to matter for cost, context-budgeting, and routing decisions. What has been missing is a peer-citable, openly distributed dataset that quantifies this drift rigorously across providers, formats, and time. LLM Tokens Atlas fills that gap: instead of a single-point anecdote, it publishes the joint distribution of offline-vs-empirical deltas with full reproducibility metadata. The cited prior observations are credited in Citation Information / Related Work below.

Source Data

Initial Data Collection and Normalization

Prompts are sampled from already-public, redistribution-friendly corpora and re-rendered into each of the five surface formats. Per-row provenance — which corpus a prompt originated from, what license it ships under, and what normalization steps were applied — is documented in data/provenance.md, maintained by the sibling atlas-corpus collector and treated as the authoritative record for downstream attribution. Source corpora include:

  • LMSYS-Chat-1M (open sample) — real chat prompts.
  • HumanEval — programming task descriptions.
  • MT-Bench — multi-turn evaluation prompts.
  • GitHub README snippets — long-form, structured technical prose.
  • Multilingual Wikipedia — for non-English coverage where applicable.
  • ShareGPT (filtered, open subset) — additional chat coverage.

Prompts are normalized into a canonical text representation, then deterministically re-rendered into each of the five formats so that format is a controlled independent variable rather than a property of the source corpus.

Who are the source language producers?

The prompt text comes from public chat logs, open-source code documentation, benchmark authors, and Wikipedia contributors. No identifying information is collected, retained, or republished beyond what is already in the source corpora; per-source licensing terms are preserved as required.

Annotations

Annotation Process

There are no human annotations. Each row records two machine-measured token counts:

  1. Offline count — computed locally by invoking the offline tokenizer shipped or reverse-engineered for the target provider/model. The exact tokenizer (with version pin) is recorded in tokenizer_id and in data/lockfile.json.
  2. Empirical count — obtained by calling the provider's authoritative token-counting endpoint or OSS tokenizer. The API version used is recorded in api_version and in data/lockfile.json.

Empirical-source coverage in v0.1.0:

  • Anthropicmessages.countTokens (HTTP, official API). Network sweep executed; empirical counts are authoritative.
  • OpenAItiktoken.encoding_for_model("gpt-4o") with o200k_base vocab, run locally. Treated as the oracle (tiktoken-as-truth) since no separate count-tokens HTTP endpoint exists.
  • Mistralmistral-tokenizer-js (vendor OSS tokenizer), run locally.

Not executed in v0.1.0 (schema-validated only, populate in v0.2.0):

  • Googlemodel.countTokens HTTP endpoint integration is validated against the schema but the empirical sweep has not been run; no Google rows are present in data/processed/atlas.parquet in v0.1.0.
  • CoherePOST /v1/tokenize integration is similarly validated but unrun; no Cohere rows are present in v0.1.0.

Both Google and Cohere sweeps land in v0.2.0 alongside their API keys (GOOGLE_API_KEY, COHERE_API_KEY); the row schema is unchanged.

Who are the annotators?

Not applicable — counts are produced by deterministic code, not annotators.

Versioning Policy

Releases use semantic versioning (MAJOR.MINOR.PATCH). Each release pins the offline tokenizer versions and provider API versions in data/lockfile.json. Because providers update their tokenizers periodically (see Considerations), a row's tokenizer_id and api_version should be treated as part of its identity for longitudinal analysis. Major-version bumps may break row schemas; minor and patch versions preserve the published schema.

Personal and Sensitive Information

The dataset does not contain new personal information beyond what is already present in the source corpora. The LMSYS-Chat-1M and ShareGPT-derived samples inherit upstream filtering. We do not collect, store, or republish any provider-side outputs other than integer token counts.

Considerations for Using the Data

Social Impact of Dataset

The intended impact is positive: more accurate cost estimation reduces wasted spend on LLM APIs, particularly for resource-constrained teams that rely on free or low-tier offline tokenizers. Better calibration also reduces the incentive for vendors to obscure tokenizer behavior, since any drift is now publicly measurable.

Discussion of Biases

  • English-heavy. Prompts disproportionately reflect English LMSYS-domain chat. Calibration deltas measured here may not generalize cleanly to predominantly non-Latin scripts where tokenizer behavior differs sharply.
  • LMSYS-domain skew. LMSYS-Chat-1M is biased toward arena-style head-to-head prompts; readers should not assume the prompt mix mirrors production traffic for an arbitrary application.
  • Provider tokenizer drift over time. Providers update their tokenizers; a calibration delta observed today may not hold next quarter. The tokenizer_id and api_version fields plus data/lockfile.json are the guard against silently comparing apples to oranges across releases.
  • Format-mediated bias. Some formats (JSON, XML) introduce structural characters that some tokenizers fuse and others split; this is the central phenomenon the dataset measures, so we record it explicitly rather than smooth it away.

Limitations

  • No generative outputs. The dataset only measures tokenization, not model quality. Cost-aware routing systems built on this data must source quality signals separately.
  • Empirical-count availability is uneven. Where no public count-tokens endpoint exists, empirical_count is null and the row participates in offline-coverage analyses only.
  • Sampling is finite. 10k-prompt class size is enough for stable per-provider deltas but lighter for some (provider × model × format × language) cells; the row counts per cell are reported in data/provenance.md.

Additional Information

Dataset Curators

Curated and maintained by Faraazuddin Mohammed (https://github.com/faraa2m). Issues, PRs, and provenance corrections welcome at https://github.com/faraa2m/llm-tokens-atlas/issues.

Licensing Information

Suggested attribution string:

"Data from llm-tokens-atlas (Faraazuddin Mohammed, 2026), https://huggingface.co/datasets/faraa2m/llm-tokens-atlas, CC-BY-4.0."

Citation Information

@misc{llm-tokens-atlas-2026,
  author       = {Faraazuddin Mohammed},
  title        = {{llm-tokens-atlas}: An Open Benchmark of LLM Tokenization Calibration Across Providers and Formats},
  year         = {2026},
  publisher    = {Hugging Face},
  howpublished = {\url{https://huggingface.co/datasets/faraa2m/llm-tokens-atlas}},
  note         = {Companion arXiv preprint forthcoming.}
}

Related Work

The offline-vs-empirical tokenizer drift phenomenon was independently surfaced in 2026 by community blog posts; this dataset is intended as the peer-citable, reproducible version of that observation, not as the discovery of it. Honest positioning credits the prior surfacers:

Methodologically related academic work:

  • Sachan et al., RouterBench: A Benchmark for Multi-LLM Routing Systems (2024) — arXiv:2403.12031. Sibling open dataset of API outcomes, but routing-focused rather than tokenization-focused.
  • Zheng et al., LMSYS-Chat-1M: A Large-Scale Real-World LLM Conversation Dataset (2023) — arXiv:2309.11998. Source corpus.
  • Ali et al., Tokenizer Choice For LLM Training: Negligible or Crucial? (2023) — arXiv:2310.08754. Training-time tokenizer study; complementary to this deployment-time drift study.
  • Rust et al., How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models (2021, ACL). Methodology precedent.
  • Javier Rando, The Worst (But Only) Claude 3 Tokenizer (2024) — https://javirando.com/blog/2024/claude-tokenizer/. Canonical source for the offline Claude tokenizer the community uses; cited here as the offline baseline for the Anthropic provider rows.

Foundation

This dataset extends the methodology and an early finding from tokenometer (cl100k_base underestimates claude-opus-4-7 tokens by ~62% median), generalizing the single-point measurement into a multi-provider, multi-format calibration distribution.

Contact

Dataset Card Authors

Faraazuddin Mohammed (https://github.com/faraa2m).

Downloads last month
26

Papers for faraa2m/llm-tokens-atlas