_leaderboard
stringclasses
1 value
_developer
stringclasses
559 values
_model
stringlengths
9
102
_uuid
stringlengths
36
36
schema_version
stringclasses
1 value
evaluation_id
stringlengths
35
133
retrieved_timestamp
stringlengths
13
18
source_data
stringclasses
1 value
evaluation_source_name
stringclasses
1 value
evaluation_source_type
stringclasses
1 value
source_organization_name
stringclasses
1 value
source_organization_url
null
source_organization_logo_url
null
evaluator_relationship
stringclasses
1 value
model_name
stringlengths
4
102
model_id
stringlengths
9
102
model_developer
stringclasses
559 values
model_inference_platform
stringclasses
1 value
evaluation_results
stringlengths
1.35k
1.41k
additional_details
stringclasses
660 values
HF Open LLM v2
LeroyDyer
LeroyDyer/SpydazWeb_AI_HumanAI_010_CHAT
7f53cef7-fba6-4802-93a2-b54f82a32d74
0.0.1
hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_010_CHAT/1762652579.7189271
1762652579.7189288
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/SpydazWeb_AI_HumanAI_010_CHAT
LeroyDyer/SpydazWeb_AI_HumanAI_010_CHAT
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2506948230694557}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
LeroyDyer
LeroyDyer/CheckPoint_C
a4fe370d-1722-4fdf-bf75-8416baeaba19
0.0.1
hfopenllm_v2/LeroyDyer_CheckPoint_C/1762652579.714836
1762652579.714837
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/CheckPoint_C
LeroyDyer/CheckPoint_C
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.34768968558979063}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
LeroyDyer
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA
431f8459-3c12-4260-a158-c58ec910590d
0.0.1
hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_IA/1762652579.720226
1762652579.720227
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30664858131978706}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
LeroyDyer
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA
bcd8c141-d286-4567-bb06-934e546a5c7c
0.0.1
hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_IA/1762652579.720018
1762652579.7200189
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30355124403250044}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
LeroyDyer
LeroyDyer/_Spydaz_Web_AI_AGI_R1_OmG_Math
983323f2-7caa-42cb-8838-8ea041303a70
0.0.1
hfopenllm_v2/LeroyDyer__Spydaz_Web_AI_AGI_R1_OmG_Math/1762652579.7249558
1762652579.724957
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/_Spydaz_Web_AI_AGI_R1_OmG_Math
LeroyDyer/_Spydaz_Web_AI_AGI_R1_OmG_Math
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5033112142448702}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
LeroyDyer
LeroyDyer/SpydazWeb_AI_HumanAI_001
f177b7f7-7143-4f72-9f9d-54fe2bc9797b
0.0.1
hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_001/1762652579.717986
1762652579.717987
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/SpydazWeb_AI_HumanAI_001
LeroyDyer/SpydazWeb_AI_HumanAI_001
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22516589316347294}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
LeroyDyer
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA
0b365c44-3cc2-4149-8614-7de6b6c2581d
0.0.1
hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_XA/1762652579.72064
1762652579.7206411
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.35788153211257245}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
LeroyDyer
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA
dc90b971-313a-4a76-b042-350adf37a43c
0.0.1
hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_XA/1762652579.720855
1762652579.720855
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA
LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA
LeroyDyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37976347203198624}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
avemio
avemio/GRAG-NEMO-12B-ORPO-HESSIAN-AI
45cc7b31-3f75-42f7-9b07-3cf704fd2b55
0.0.1
hfopenllm_v2/avemio_GRAG-NEMO-12B-ORPO-HESSIAN-AI/1762652580.020413
1762652580.0204139
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
avemio/GRAG-NEMO-12B-ORPO-HESSIAN-AI
avemio/GRAG-NEMO-12B-ORPO-HESSIAN-AI
avemio
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.0}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH", "lower_is_be...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-2.2-rys-78b
cfaafe4c-50a1-4cde-b092-fdbaeea86fb3
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-2.2-rys-78b/1762652579.754511
1762652579.754511
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-2.2-rys-78b
MaziyarPanahi/calme-2.2-rys-78b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7986420475449585}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Llama-3-8B-Instruct-v0.8
c68859dd-6db0-4bdc-a031-92ac7d1d2585
0.0.1
hfopenllm_v2/MaziyarPanahi_Llama-3-8B-Instruct-v0.8/1762652579.750486
1762652579.750487
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Llama-3-8B-Instruct-v0.8
MaziyarPanahi/Llama-3-8B-Instruct-v0.8
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7527549125209998}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.1-instruct-78b
898e5e91-c4c0-4494-baad-37c2bfd1931b
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.1-instruct-78b/1762652579.7584739
1762652579.758475
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.1-instruct-78b
MaziyarPanahi/calme-3.1-instruct-78b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8135547015252862}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Llama-3-8B-Instruct-v0.10
19143059-07d5-44b2-b599-193147f6196a
0.0.1
hfopenllm_v2/MaziyarPanahi_Llama-3-8B-Instruct-v0.10/1762652579.750272
1762652579.750272
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Llama-3-8B-Instruct-v0.10
MaziyarPanahi/Llama-3-8B-Instruct-v0.10
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7667433520835827}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Calme-4x7B-MoE-v0.2
ca2df1c9-79b2-453b-9cd1-b607e48f5dd7
0.0.1
hfopenllm_v2/MaziyarPanahi_Calme-4x7B-MoE-v0.2/1762652579.7498329
1762652579.749834
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Calme-4x7B-MoE-v0.2
MaziyarPanahi/Calme-4x7B-MoE-v0.2
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.429447200095746}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "MixtralForCausalLM", "params_billions": 24.154}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-2.3-rys-78b
33a06134-e58d-4bc7-8421-c5ae2f0dcd1f
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-2.3-rys-78b/1762652579.7562392
1762652579.7562408
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-2.3-rys-78b
MaziyarPanahi/calme-2.3-rys-78b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8065854155862002}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.2-instruct-3b
83e46bac-5266-4f65-a4dd-76240b297adc
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.2-instruct-3b/1762652579.759095
1762652579.7590961
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.2-instruct-3b
MaziyarPanahi/calme-3.2-instruct-3b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5533196363426819}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.3-instruct-3b
8aa85bd2-eab2-491b-95a3-ac6321cbe298
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.3-instruct-3b/1762652579.759784
1762652579.759785
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.3-instruct-3b
MaziyarPanahi/calme-3.3-instruct-3b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6423212631373645}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.2-baguette-3b
e49441f3-99a5-4cdb-bff1-79cc21711bab
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.2-baguette-3b/1762652579.75889
1762652579.758891
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.2-baguette-3b
MaziyarPanahi/calme-3.2-baguette-3b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6338282423968404}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.085}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.1-instruct-3b
67915bce-0b54-4996-90f6-cec6def9bbba
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.1-instruct-3b/1762652579.758249
1762652579.75825
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.1-instruct-3b
MaziyarPanahi/calme-3.1-instruct-3b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43359397509718656}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.085}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.1-baguette-3b
8f0a6518-d153-43ec-b426-02136a2bc367
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.1-baguette-3b/1762652579.7580318
1762652579.7580328
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.1-baguette-3b
MaziyarPanahi/calme-3.1-baguette-3b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6234369251364158}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.085}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.2-instruct-78b
77cc280c-b794-4a9a-addc-e2eb0a1af896
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.2-instruct-78b/1762652579.759298
1762652579.759299
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.2-instruct-78b
MaziyarPanahi/calme-3.2-instruct-78b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8062607215521482}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Calme-4x7B-MoE-v0.1
f4512664-c531-4b13-b76e-e96c2b03febf
0.0.1
hfopenllm_v2/MaziyarPanahi_Calme-4x7B-MoE-v0.1/1762652579.7495291
1762652579.74953
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Calme-4x7B-MoE-v0.1
MaziyarPanahi/Calme-4x7B-MoE-v0.1
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4315205875964663}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MixtralForCausalLM", "params_billions": 24.154}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Llama-3-8B-Instruct-v0.9
1fb0056b-4f66-404b-89ac-a58185747ce2
0.0.1
hfopenllm_v2/MaziyarPanahi_Llama-3-8B-Instruct-v0.9/1762652579.750697
1762652579.750697
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Llama-3-8B-Instruct-v0.9
MaziyarPanahi/Llama-3-8B-Instruct-v0.9
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.763046494412603}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-2.1-rys-78b
387000a4-7ef5-46c6-9b5e-9bfe7c2cfc18
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-2.1-rys-78b/1762652579.752971
1762652579.752971
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-2.1-rys-78b
MaziyarPanahi/calme-2.1-rys-78b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8135547015252862}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Qwen2-7B-Instruct-v0.8
a65af628-f518-4da7-afc5-7cba4234415b
0.0.1
hfopenllm_v2/MaziyarPanahi_Qwen2-7B-Instruct-v0.8/1762652579.751401
1762652579.751402
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Qwen2-7B-Instruct-v0.8
MaziyarPanahi/Qwen2-7B-Instruct-v0.8
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.27747266142723526}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-3.3-baguette-3b
22cbbb6d-1014-42af-96cf-1636fcb40679
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-3.3-baguette-3b/1762652579.759511
1762652579.759511
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-3.3-baguette-3b
MaziyarPanahi/calme-3.3-baguette-3b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6359514975819713}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/calme-2.4-rys-78b
48433dc8-40ff-4e36-8c6a-ced33bc22e4f
0.0.1
hfopenllm_v2/MaziyarPanahi_calme-2.4-rys-78b/1762652579.7570088
1762652579.75701
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/calme-2.4-rys-78b
MaziyarPanahi/calme-2.4-rys-78b
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8010899967641414}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Qwen2-7B-Instruct-v0.1
ce4ee4fe-8a38-467b-b189-b25311c23c4e
0.0.1
hfopenllm_v2/MaziyarPanahi_Qwen2-7B-Instruct-v0.1/1762652579.7511811
1762652579.751182
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Qwen2-7B-Instruct-v0.1
MaziyarPanahi/Qwen2-7B-Instruct-v0.1
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.33522498082864577}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
MaziyarPanahi
MaziyarPanahi/Llama-3-70B-Instruct-v0.1
1e2759fa-3e87-447b-b0ca-5a4e2e293589
0.0.1
hfopenllm_v2/MaziyarPanahi_Llama-3-70B-Instruct-v0.1/1762652579.750048
1762652579.750049
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
MaziyarPanahi/Llama-3-70B-Instruct-v0.1
MaziyarPanahi/Llama-3-70B-Instruct-v0.1
MaziyarPanahi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47143800671108216}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 70.554}
HF Open LLM v2
PJMixers-Dev
PJMixers-Dev/LLaMa-3.1-Instruct-Interleaved-Zeroed-13B
fb66b283-bfd6-4437-95b7-d74a0d8d2814
0.0.1
hfopenllm_v2/PJMixers-Dev_LLaMa-3.1-Instruct-Interleaved-Zeroed-13B/1762652579.809847
1762652579.809848
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PJMixers-Dev/LLaMa-3.1-Instruct-Interleaved-Zeroed-13B
PJMixers-Dev/LLaMa-3.1-Instruct-Interleaved-Zeroed-13B
PJMixers-Dev
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7871015572015585}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 13.047}
HF Open LLM v2
PJMixers-Dev
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
1d91cdce-0bdb-4567-9296-6225db3aa0bc
0.0.1
hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B/1762652579.8105159
1762652579.810517
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
PJMixers-Dev
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.693054428915278}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
PJMixers-Dev
PJMixers-Dev/LLaMa-3.2-Instruct-JankMixBread-v0.1-3B
56f36430-4bb1-425d-ac4b-30d85237667c
0.0.1
hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMixBread-v0.1-3B/1762652579.8111491
1762652579.81115
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PJMixers-Dev/LLaMa-3.2-Instruct-JankMixBread-v0.1-3B
PJMixers-Dev/LLaMa-3.2-Instruct-JankMixBread-v0.1-3B
PJMixers-Dev
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5040858256093831}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
PJMixers-Dev
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B
d1875dfd-05ab-4a49-8c7f-02cddf35a695
0.0.1
hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B/1762652579.810729
1762652579.81073
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B
PJMixers-Dev
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6291573026237051}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
PJMixers-Dev
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B
62b12d95-1da2-407c-8552-8c5e951c5c85
0.0.1
hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B/1762652579.8109388
1762652579.8109398
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B
PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B
PJMixers-Dev
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6503898544750152}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
PJMixers-Dev
PJMixers-Dev/L3.2-Instruct-Thinking-v0.1-1B
3c942d2f-0b53-498e-ab05-71d5075cb974
0.0.1
hfopenllm_v2/PJMixers-Dev_L3.2-Instruct-Thinking-v0.1-1B/1762652579.8095942
1762652579.8095949
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PJMixers-Dev/L3.2-Instruct-Thinking-v0.1-1B
PJMixers-Dev/L3.2-Instruct-Thinking-v0.1-1B
PJMixers-Dev
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46276989498973836}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.236}
HF Open LLM v2
AicoresSecurity
AicoresSecurity/Cybernet-Sec-3B-R1-V1
b613ecbe-7b2b-4b03-ab2c-163f9988a8fc
0.0.1
hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V1/1762652579.478252
1762652579.4782531
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
AicoresSecurity/Cybernet-Sec-3B-R1-V1
AicoresSecurity/Cybernet-Sec-3B-R1-V1
AicoresSecurity
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6145693426774292}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
AicoresSecurity
AicoresSecurity/Cybernet-Sec-3B-R1-V0
38f169f0-e939-4b12-8f78-b2a27fb90de0
0.0.1
hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V0/1762652579.4777558
1762652579.477757
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
AicoresSecurity/Cybernet-Sec-3B-R1-V0
AicoresSecurity/Cybernet-Sec-3B-R1-V0
AicoresSecurity
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6358018945287394}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
AicoresSecurity
AicoresSecurity/Cybernet-Sec-3B-R1-V1.1
e8c63728-a1f5-432f-bf9f-204b0f4041aa
0.0.1
hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V1.1/1762652579.478466
1762652579.478467
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
AicoresSecurity/Cybernet-Sec-3B-R1-V1.1
AicoresSecurity/Cybernet-Sec-3B-R1-V1.1
AicoresSecurity
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6730209178313542}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
AicoresSecurity
AicoresSecurity/Cybernet-Sec-3B-R1-V0-Coder
48732edf-8baf-438e-8a5c-763eee6c0c18
0.0.1
hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V0-Coder/1762652579.478028
1762652579.478029
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
AicoresSecurity/Cybernet-Sec-3B-R1-V0-Coder
AicoresSecurity/Cybernet-Sec-3B-R1-V0-Coder
AicoresSecurity
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7097656440466851}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
tiiuae
tiiuae/falcon-40b
cfdece82-631e-48b7-8232-91a8d9ccf65c
0.0.1
hfopenllm_v2/tiiuae_falcon-40b/1762652580.568969
1762652580.56897
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/falcon-40b
tiiuae/falcon-40b
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24964538535530173}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 40.0}
HF Open LLM v2
tiiuae
tiiuae/falcon-11B
705a1ff4-2e40-4827-af54-099870fac588
0.0.1
hfopenllm_v2/tiiuae_falcon-11B/1762652580.568774
1762652580.568774
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/falcon-11B
tiiuae/falcon-11B
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3261324397044287}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 11.103}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-3B-Base
1b0d1ae7-322b-46d2-bc33-160f578499b1
0.0.1
hfopenllm_v2/tiiuae_Falcon3-3B-Base/1762652580.5675461
1762652580.5675468
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-3B-Base
tiiuae/Falcon3-3B-Base
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2764985793250797}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.228}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-10B-Instruct
741838df-e2a3-4c54-84d3-fe491444071b
0.0.1
hfopenllm_v2/tiiuae_Falcon3-10B-Instruct/1762652580.566902
1762652580.566903
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-10B-Instruct
tiiuae/Falcon3-10B-Instruct
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7816560060639104}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-7B-Instruct
ed988bd0-76b0-4ab6-9c9e-5a5e0aefb936
0.0.1
hfopenllm_v2/tiiuae_Falcon3-7B-Instruct/1762652580.568164
1762652580.568164
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-7B-Instruct
tiiuae/Falcon3-7B-Instruct
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7612479332615238}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-1B-Base
1e11a625-87e1-49d0-94a6-8f9ec1f75fc3
0.0.1
hfopenllm_v2/tiiuae_Falcon3-1B-Base/1762652580.567122
1762652580.567122
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-1B-Base
tiiuae/Falcon3-1B-Base
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24280132271262472}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.669}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-1B-Instruct
a060e2b0-d1ae-48b7-b8f9-c51fadc3e152
0.0.1
hfopenllm_v2/tiiuae_Falcon3-1B-Instruct/1762652580.567335
1762652580.567335
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-1B-Instruct
tiiuae/Falcon3-1B-Instruct
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5556678501930433}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.669}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-10B-Base
4e1ce0d3-f454-480b-a4f7-7aa827eaaf1a
0.0.1
hfopenllm_v2/tiiuae_Falcon3-10B-Base/1762652580.566659
1762652580.566659
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-10B-Base
tiiuae/Falcon3-10B-Base
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3647754624396601}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306}
HF Open LLM v2
tiiuae
tiiuae/falcon-40b-instruct
1d6f8802-e9aa-471c-8fbc-1cd807357ab5
0.0.1
hfopenllm_v2/tiiuae_falcon-40b-instruct/1762652580.569173
1762652580.569173
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/falcon-40b-instruct
tiiuae/falcon-40b-instruct
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24544874266945038}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 40.0}
HF Open LLM v2
tiiuae
tiiuae/falcon-7b
0e9837cb-4dda-4058-a89e-4127b5980eed
0.0.1
hfopenllm_v2/tiiuae_falcon-7b/1762652580.5693781
1762652580.569379
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/falcon-7b
tiiuae/falcon-7b
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.182051401392749}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 7.0}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-Mamba-7B-Base
766e6e63-5779-49cd-9e8c-2bc475c1356a
0.0.1
hfopenllm_v2/tiiuae_Falcon3-Mamba-7B-Base/1762652580.568367
1762652580.5683682
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-Mamba-7B-Base
tiiuae/Falcon3-Mamba-7B-Base
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28911288713945665}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "FalconMambaForCausalLM", "params_billions": 7.273}
HF Open LLM v2
tiiuae
tiiuae/falcon-mamba-7b
9878c419-fff8-402a-a315-70864e5ae60c
0.0.1
hfopenllm_v2/tiiuae_falcon-mamba-7b/1762652580.569833
1762652580.569834
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/falcon-mamba-7b
tiiuae/falcon-mamba-7b
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3335760227307987}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "FalconMambaForCausalLM", "params_billions": 7.0}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-3B-Instruct
7aa3aa0e-3b5e-4c0c-a697-2e87859c44f2
0.0.1
hfopenllm_v2/tiiuae_Falcon3-3B-Instruct/1762652580.567748
1762652580.567749
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-3B-Instruct
tiiuae/Falcon3-3B-Instruct
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6976755010040027}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.228}
HF Open LLM v2
tiiuae
tiiuae/falcon-7b-instruct
2b84722f-58fc-421d-ae1a-9e21ac0b4080
0.0.1
hfopenllm_v2/tiiuae_falcon-7b-instruct/1762652580.5696268
1762652580.5696268
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/falcon-7b-instruct
tiiuae/falcon-7b-instruct
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19688869976107837}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 7.0}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-7B-Base
2420519c-81f1-43b3-9b76-af141d2574f4
0.0.1
hfopenllm_v2/tiiuae_Falcon3-7B-Base/1762652580.56796
1762652580.567961
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-7B-Base
tiiuae/Falcon3-7B-Base
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.34159474638403875}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
tiiuae
tiiuae/Falcon3-Mamba-7B-Instruct
69491efc-0287-4288-bdf0-bcc57c53b94e
0.0.1
hfopenllm_v2/tiiuae_Falcon3-Mamba-7B-Instruct/1762652580.5685718
1762652580.5685718
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
tiiuae/Falcon3-Mamba-7B-Instruct
tiiuae/Falcon3-Mamba-7B-Instruct
tiiuae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7165099713205406}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "FalconMambaForCausalLM", "params_billions": 7.273}
HF Open LLM v2
alpindale
alpindale/WizardLM-2-8x22B
c2899c4e-5bc9-4b0b-8938-b9848b86fe37
0.0.1
hfopenllm_v2/alpindale_WizardLM-2-8x22B/1762652580.009551
1762652580.0095518
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
alpindale/WizardLM-2-8x22B
alpindale/WizardLM-2-8x22B
alpindale
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5272166739805937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MixtralForCausalLM", "params_billions": 140.621}
HF Open LLM v2
alpindale
alpindale/magnum-72b-v1
186687f8-ed25-44c9-b634-36db1c734844
0.0.1
hfopenllm_v2/alpindale_magnum-72b-v1/1762652580.0098088
1762652580.00981
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
alpindale/magnum-72b-v1
alpindale/magnum-72b-v1
alpindale
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7606484128778308}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706}
HF Open LLM v2
Amaorynho
Amaorynho/BBAI_375
ad4b6e40-883c-47c5-ba33-6c112c2c6b09
0.0.1
hfopenllm_v2/Amaorynho_BBAI_375/1762652579.480799
1762652579.480799
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Amaorynho/BBAI_375
Amaorynho/BBAI_375
Amaorynho
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14670518668244703}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.09}
HF Open LLM v2
Amaorynho
Amaorynho/BBAI270V4
183313de-d526-42a9-a35d-a4e71466e546
0.0.1
hfopenllm_v2/Amaorynho_BBAI270V4/1762652579.4803882
1762652579.4803882
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Amaorynho/BBAI270V4
Amaorynho/BBAI270V4
Amaorynho
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1990374428737971}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
Amaorynho
Amaorynho/BBAIIFEV1
7c0342a3-5bd4-47b0-b238-d5dcb0f6236e
0.0.1
hfopenllm_v2/Amaorynho_BBAIIFEV1/1762652579.480599
1762652579.4806
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Amaorynho/BBAIIFEV1
Amaorynho/BBAIIFEV1
Amaorynho
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8047369867507104}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
Amaorynho
Amaorynho/BBAI2006
ef37c096-a089-4d3e-9fad-c0f959a18bb3
0.0.1
hfopenllm_v2/Amaorynho_BBAI2006/1762652579.480136
1762652579.4801369
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Amaorynho/BBAI2006
Amaorynho/BBAI2006
Amaorynho
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14670518668244703}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.09}
HF Open LLM v2
NikolaSigmoid
NikolaSigmoid/AceMath-1.5B-Instruct-1epoch
0cf3db2f-9b23-4602-ac92-265bafd36410
0.0.1
hfopenllm_v2/NikolaSigmoid_AceMath-1.5B-Instruct-1epoch/1762652579.783191
1762652579.7831922
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
NikolaSigmoid/AceMath-1.5B-Instruct-1epoch
NikolaSigmoid/AceMath-1.5B-Instruct-1epoch
NikolaSigmoid
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2848918646967823}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.791}
HF Open LLM v2
NikolaSigmoid
NikolaSigmoid/acemath-200
4414a96e-0664-4531-9c0f-3eb4a062fbe2
0.0.1
hfopenllm_v2/NikolaSigmoid_acemath-200/1762652579.783974
1762652579.783974
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
NikolaSigmoid/acemath-200
NikolaSigmoid/acemath-200
NikolaSigmoid
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2848918646967823}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.791}
HF Open LLM v2
NikolaSigmoid
NikolaSigmoid/AceMath-1.5B-Instruct-dolphin-r1-200
93f56942-30d8-4a0f-af8d-901fb264436c
0.0.1
hfopenllm_v2/NikolaSigmoid_AceMath-1.5B-Instruct-dolphin-r1-200/1762652579.783446
1762652579.783447
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
NikolaSigmoid/AceMath-1.5B-Instruct-dolphin-r1-200
NikolaSigmoid/AceMath-1.5B-Instruct-dolphin-r1-200
NikolaSigmoid
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.18080249294095221}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.928}
HF Open LLM v2
Alepach
Alepach/notHumpback-M1-v2
27c6c36d-6bd5-439b-bdc8-1bd0f8f4c9ea
0.0.1
hfopenllm_v2/Alepach_notHumpback-M1-v2/1762652579.4791439
1762652579.479145
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Alepach/notHumpback-M1-v2
Alepach/notHumpback-M1-v2
Alepach
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2277135777514772}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
Alepach
Alepach/notHumpback-M1
030f17b0-036f-4021-90da-6c1d38da659d
0.0.1
hfopenllm_v2/Alepach_notHumpback-M1/1762652579.478936
1762652579.4789371
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Alepach/notHumpback-M1
Alepach/notHumpback-M1
Alepach
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2206944241279804}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
Alepach
Alepach/notHumpback-M0
1a4477f7-c414-41ab-bbcb-593f4a86031a
0.0.1
hfopenllm_v2/Alepach_notHumpback-M0/1762652579.4786859
1762652579.478687
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Alepach/notHumpback-M0
Alepach/notHumpback-M0
Alepach
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23500755772461512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
win10
win10/miscii-14b-1M-0128
c19f2ddd-7710-4844-9f1f-c0cd1c7e3e41
0.0.1
hfopenllm_v2/win10_miscii-14b-1M-0128/1762652580.5956988
1762652580.5957
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
win10/miscii-14b-1M-0128
win10/miscii-14b-1M-0128
win10
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4180818007331658}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
win10
win10/Llama-3.2-3B-Instruct-24-9-29
bf253a63-4685-4e51-8a0d-5209306926c8
0.0.1
hfopenllm_v2/win10_Llama-3.2-3B-Instruct-24-9-29/1762652580.594629
1762652580.59463
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
win10/Llama-3.2-3B-Instruct-24-9-29
win10/Llama-3.2-3B-Instruct-24-9-29
win10
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7332211864519476}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
win10
win10/Breeze-13B-32k-Instruct-v1_0
bc990db1-c6d9-4113-9946-466bfd5cf9cc
0.0.1
hfopenllm_v2/win10_Breeze-13B-32k-Instruct-v1_0/1762652580.5941818
1762652580.594183
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
win10/Breeze-13B-32k-Instruct-v1_0
win10/Breeze-13B-32k-Instruct-v1_0
win10
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.35843118481185476}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.726}
HF Open LLM v2
win10
win10/llama3-13.45b-Instruct
3c9eb291-6171-4d40-aa5f-58d39738fdcb
0.0.1
hfopenllm_v2/win10_llama3-13.45b-Instruct/1762652580.595499
1762652580.5955
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
win10/llama3-13.45b-Instruct
win10/llama3-13.45b-Instruct
win10
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4144348107465968}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 13.265}
HF Open LLM v2
win10
win10/Qwen2.5-2B-Instruct
143dc973-1063-45d6-9747-9f24a9ae5657
0.0.1
hfopenllm_v2/win10_Qwen2.5-2B-Instruct/1762652580.5952861
1762652580.595287
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
win10/Qwen2.5-2B-Instruct
win10/Qwen2.5-2B-Instruct
win10
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22728914834860392}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 2.9}
HF Open LLM v2
win10
win10/ArliAI-RPMax-v1.3-merge-13.3B
16777b0f-3063-45eb-be07-294d13f975ac
0.0.1
hfopenllm_v2/win10_ArliAI-RPMax-v1.3-merge-13.3B/1762652580.593927
1762652580.5939279
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
win10/ArliAI-RPMax-v1.3-merge-13.3B
win10/ArliAI-RPMax-v1.3-merge-13.3B
win10
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3038260703821416}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 13.265}
HF Open LLM v2
bigscience
bigscience/bloom-3b
88f90805-7410-4ec1-ad19-8e8a146f1ba3
0.0.1
hfopenllm_v2/bigscience_bloom-3b/1762652580.034177
1762652580.034179
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
bigscience/bloom-3b
bigscience/bloom-3b
bigscience
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1270961050013963}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 3.003}
HF Open LLM v2
bigscience
bigscience/bloom-1b1
284ba4fb-cae4-46ac-a5dd-a36fb145da55
0.0.1
hfopenllm_v2/bigscience_bloom-1b1/1762652580.033589
1762652580.033589
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
bigscience/bloom-1b1
bigscience/bloom-1b1
bigscience
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13733781920858879}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 1.065}
HF Open LLM v2
bigscience
bigscience/bloom-560m
82454b92-cca1-4ac8-a620-e1a8487a5b8e
0.0.1
hfopenllm_v2/bigscience_bloom-560m/1762652580.034546
1762652580.034548
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
bigscience/bloom-560m
bigscience/bloom-560m
bigscience
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.06202431769926019}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 0.559}
HF Open LLM v2
bigscience
bigscience/bloom-1b7
8adb8bb9-d057-45df-827a-cd8f014b4ff6
0.0.1
hfopenllm_v2/bigscience_bloom-1b7/1762652580.033839
1762652580.033839
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
bigscience/bloom-1b7
bigscience/bloom-1b7
bigscience
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.10438968603305895}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 1.722}
HF Open LLM v2
bigscience
bigscience/bloom-7b1
d5fe1452-b6ee-4f1d-9eca-713b49a6a941
0.0.1
hfopenllm_v2/bigscience_bloom-7b1/1762652580.0348449
1762652580.034846
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
bigscience/bloom-7b1
bigscience/bloom-7b1
bigscience
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13221696210499254}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "BloomForCausalLM", "params_billions": 7.069}
HF Open LLM v2
yuvraj17
yuvraj17/Llama3-8B-SuperNova-Spectrum-Hermes-DPO
d22c83a1-9c1c-43df-b033-c6cb75cb389d
0.0.1
hfopenllm_v2/yuvraj17_Llama3-8B-SuperNova-Spectrum-Hermes-DPO/1762652580.611586
1762652580.611586
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
yuvraj17/Llama3-8B-SuperNova-Spectrum-Hermes-DPO
yuvraj17/Llama3-8B-SuperNova-Spectrum-Hermes-DPO
yuvraj17
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4690897928607206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
hon9kon9ize
hon9kon9ize/CantoneseLLMChat-v1.0-7B
cccf983e-e1b8-4f0f-b147-abccdea65548
0.0.1
hfopenllm_v2/hon9kon9ize_CantoneseLLMChat-v1.0-7B/1762652580.191013
1762652580.191013
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hon9kon9ize/CantoneseLLMChat-v1.0-7B
hon9kon9ize/CantoneseLLMChat-v1.0-7B
hon9kon9ize
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44548353923146145}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
hon9kon9ize
hon9kon9ize/CantoneseLLMChat-v0.5
6e87be06-ca0e-48a4-ae28-4a5794600117
0.0.1
hfopenllm_v2/hon9kon9ize_CantoneseLLMChat-v0.5/1762652580.190754
1762652580.1907551
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hon9kon9ize/CantoneseLLMChat-v0.5
hon9kon9ize/CantoneseLLMChat-v0.5
hon9kon9ize
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3230849701015528}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 6.069}
HF Open LLM v2
Khetterman
Khetterman/DarkAtom-12B-v3
64802b86-879e-4072-b5ad-aab17d7251f0
0.0.1
hfopenllm_v2/Khetterman_DarkAtom-12B-v3/1762652579.6987362
1762652579.698737
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Khetterman/DarkAtom-12B-v3
Khetterman/DarkAtom-12B-v3
Khetterman
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6173419859306639}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
Khetterman
Khetterman/Kosmos-8B-v1
936cbaa1-e55b-46b8-9610-a5a8faaf4434
0.0.1
hfopenllm_v2/Khetterman_Kosmos-8B-v1/1762652579.6990001
1762652579.699001
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Khetterman/Kosmos-8B-v1
Khetterman/Kosmos-8B-v1
Khetterman
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41291107594515886}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
RezVortex
RezVortex/JAJUKA-WEWILLNEVERFORGETYOU-3B
76f26fef-fa87-4cf5-a317-ea4b743e7432
0.0.1
hfopenllm_v2/RezVortex_JAJUKA-WEWILLNEVERFORGETYOU-3B/1762652579.853197
1762652579.853197
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
RezVortex/JAJUKA-WEWILLNEVERFORGETYOU-3B
RezVortex/JAJUKA-WEWILLNEVERFORGETYOU-3B
RezVortex
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6858103166265509}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
RezVortex
RezVortex/Jajuka-3b
a41d111c-dd5d-4f77-b52d-9a2dc9f31e50
0.0.1
hfopenllm_v2/RezVortex_Jajuka-3b/1762652579.85344
1762652579.853441
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
RezVortex/Jajuka-3b
RezVortex/Jajuka-3b
RezVortex
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6925047762159957}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
formulae
formulae/mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp
936751f5-4483-4986-9a8c-cb002feb8858
0.0.1
hfopenllm_v2/formulae_mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp/1762652580.1578538
1762652580.157855
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp
formulae/mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.16139288199754429}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-elite-v1.1-gen2-7b-2-25-2025
106c33d2-84fb-4ea3-b2d3-78981834fdb0
0.0.1
hfopenllm_v2/formulae_mita-elite-v1.1-gen2-7b-2-25-2025/1762652580.158336
1762652580.158336
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-elite-v1.1-gen2-7b-2-25-2025
formulae/mita-elite-v1.1-gen2-7b-2-25-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14108454456397912}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-v1.2-7b-2-24-2025
a07149d4-66e5-4a0d-b4ae-b696027e821c
0.0.1
hfopenllm_v2/formulae_mita-v1.2-7b-2-24-2025/1762652580.160727
1762652580.160728
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-v1.2-7b-2-24-2025
formulae/mita-v1.2-7b-2-24-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.256415200556745}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-elite-v1.2-7b-2-26-2025
761560dc-3a0b-481f-8ec2-4d1ea97cfa6f
0.0.1
hfopenllm_v2/formulae_mita-elite-v1.2-7b-2-26-2025/1762652580.158752
1762652580.158756
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-elite-v1.2-7b-2-26-2025
formulae/mita-elite-v1.2-7b-2-26-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14800396281865452}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-v1.1-7b-2-24-2025
332cbdd8-96b7-40d5-87c6-3610dcbcdc54
0.0.1
hfopenllm_v2/formulae_mita-v1.1-7b-2-24-2025/1762652580.1604211
1762652580.1604218
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-v1.1-7b-2-24-2025
formulae/mita-v1.1-7b-2-24-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.34122018466557624}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-gen3-v1.2-7b-2-26-2025
a28f8779-d2df-4371-b946-472b335f3ca3
0.0.1
hfopenllm_v2/formulae_mita-gen3-v1.2-7b-2-26-2025/1762652580.15945
1762652580.1594508
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-gen3-v1.2-7b-2-26-2025
formulae/mita-gen3-v1.2-7b-2-26-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2043577707150361}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-gen3-7b-2-26-2025
0aa40e02-762d-4a80-932f-f967057c4f50
0.0.1
hfopenllm_v2/formulae_mita-gen3-7b-2-26-2025/1762652580.159164
1762652580.159165
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-gen3-7b-2-26-2025
formulae/mita-gen3-7b-2-26-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1964144026737944}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-math-v2.3-2-25-2025
fa005333-c7b5-4494-a8cb-4edb1f7d00b9
0.0.1
hfopenllm_v2/formulae_mita-math-v2.3-2-25-2025/1762652580.159737
1762652580.159738
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-math-v2.3-2-25-2025
formulae/mita-math-v2.3-2-25-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13733781920858879}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-v1-7b
9c629542-6fd0-4cd1-90c7-7f1e95a7a25e
0.0.1
hfopenllm_v2/formulae_mita-v1-7b/1762652580.160087
1762652580.160088
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-v1-7b
formulae/mita-v1-7b
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19723888172271792}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
formulae
formulae/mita-elite-v1.1-7b-2-25-2025
7352f47c-8b57-477f-8190-b08b5b23dfb5
0.0.1
hfopenllm_v2/formulae_mita-elite-v1.1-7b-2-25-2025/1762652580.158112
1762652580.158113
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
formulae/mita-elite-v1.1-7b-2-25-2025
formulae/mita-elite-v1.1-7b-2-25-2025
formulae
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1249728498162653}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
sometimesanotion
sometimesanotion/lamarck-14b-prose-model_stock
3191b3a3-761a-42b4-bd31-b8dc22a4c722
0.0.1
hfopenllm_v2/sometimesanotion_lamarck-14b-prose-model_stock/1762652580.5312169
1762652580.5312169
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sometimesanotion/lamarck-14b-prose-model_stock
sometimesanotion/lamarck-14b-prose-model_stock
sometimesanotion
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4276486389446668}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
sometimesanotion
sometimesanotion/Lamarck-14B-v0.6-002-model_stock
bd904778-1ad9-48fe-a12e-4b62ce46bd0b
0.0.1
hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.6-002-model_stock/1762652580.520087
1762652580.520087
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sometimesanotion/Lamarck-14B-v0.6-002-model_stock
sometimesanotion/Lamarck-14B-v0.6-002-model_stock
sometimesanotion
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.669224324791553}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.0}
HF Open LLM v2
sometimesanotion
sometimesanotion/Lamarck-14B-v0.7-rc1
5919f71f-8d7b-4cce-a7ce-01680c08acf2
0.0.1
hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.7-rc1/1762652580.520714
1762652580.520715
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sometimesanotion/Lamarck-14B-v0.7-rc1
sometimesanotion/Lamarck-14B-v0.7-rc1
sometimesanotion
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7305482785675341}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
sometimesanotion
sometimesanotion/Lamarck-14B-v0.3
6103d107-0eb8-4b0e-8947-d5c7e7cb62f6
0.0.1
hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.3/1762652580.519407
1762652580.5194082
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sometimesanotion/Lamarck-14B-v0.3
sometimesanotion/Lamarck-14B-v0.3
sometimesanotion
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5031616111916382}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
sometimesanotion
sometimesanotion/Lamarck-14B-v0.1-experimental
aa2b9fb3-77ca-4a48-b3dd-77879220a6b8
0.0.1
hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.1-experimental/1762652580.519198
1762652580.519199
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sometimesanotion/Lamarck-14B-v0.1-experimental
sometimesanotion/Lamarck-14B-v0.1-experimental
sometimesanotion
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5353850006870658}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}