_leaderboard stringclasses 1
value | _developer stringclasses 559
values | _model stringlengths 9 102 | _uuid stringlengths 36 36 | schema_version stringclasses 1
value | evaluation_id stringlengths 35 133 | retrieved_timestamp stringlengths 13 18 | source_data stringclasses 1
value | evaluation_source_name stringclasses 1
value | evaluation_source_type stringclasses 1
value | source_organization_name stringclasses 1
value | source_organization_url null | source_organization_logo_url null | evaluator_relationship stringclasses 1
value | model_name stringlengths 4 102 | model_id stringlengths 9 102 | model_developer stringclasses 559
values | model_inference_platform stringclasses 1
value | evaluation_results stringlengths 1.35k 1.41k | additional_details stringclasses 660
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HF Open LLM v2 | LeroyDyer | LeroyDyer/SpydazWeb_AI_HumanAI_010_CHAT | 7f53cef7-fba6-4802-93a2-b54f82a32d74 | 0.0.1 | hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_010_CHAT/1762652579.7189271 | 1762652579.7189288 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/SpydazWeb_AI_HumanAI_010_CHAT | LeroyDyer/SpydazWeb_AI_HumanAI_010_CHAT | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2506948230694557}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | LeroyDyer | LeroyDyer/CheckPoint_C | a4fe370d-1722-4fdf-bf75-8416baeaba19 | 0.0.1 | hfopenllm_v2/LeroyDyer_CheckPoint_C/1762652579.714836 | 1762652579.714837 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/CheckPoint_C | LeroyDyer/CheckPoint_C | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.34768968558979063}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | LeroyDyer | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA | 431f8459-3c12-4260-a158-c58ec910590d | 0.0.1 | hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_IA/1762652579.720226 | 1762652579.720227 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30664858131978706}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | LeroyDyer | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA | bcd8c141-d286-4567-bb06-934e546a5c7c | 0.0.1 | hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_IA/1762652579.720018 | 1762652579.7200189 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_IA | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30355124403250044}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | LeroyDyer | LeroyDyer/_Spydaz_Web_AI_AGI_R1_OmG_Math | 983323f2-7caa-42cb-8838-8ea041303a70 | 0.0.1 | hfopenllm_v2/LeroyDyer__Spydaz_Web_AI_AGI_R1_OmG_Math/1762652579.7249558 | 1762652579.724957 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/_Spydaz_Web_AI_AGI_R1_OmG_Math | LeroyDyer/_Spydaz_Web_AI_AGI_R1_OmG_Math | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5033112142448702}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | LeroyDyer | LeroyDyer/SpydazWeb_AI_HumanAI_001 | f177b7f7-7143-4f72-9f9d-54fe2bc9797b | 0.0.1 | hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_001/1762652579.717986 | 1762652579.717987 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/SpydazWeb_AI_HumanAI_001 | LeroyDyer/SpydazWeb_AI_HumanAI_001 | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22516589316347294}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | LeroyDyer | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA | 0b365c44-3cc2-4149-8614-7de6b6c2581d | 0.0.1 | hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_XA/1762652579.72064 | 1762652579.7206411 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.35788153211257245}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | LeroyDyer | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA | dc90b971-313a-4a76-b042-350adf37a43c | 0.0.1 | hfopenllm_v2/LeroyDyer_SpydazWeb_AI_HumanAI_012_INSTRUCT_XA/1762652579.720855 | 1762652579.720855 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA | LeroyDyer/SpydazWeb_AI_HumanAI_012_INSTRUCT_XA | LeroyDyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37976347203198624}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | avemio | avemio/GRAG-NEMO-12B-ORPO-HESSIAN-AI | 45cc7b31-3f75-42f7-9b07-3cf704fd2b55 | 0.0.1 | hfopenllm_v2/avemio_GRAG-NEMO-12B-ORPO-HESSIAN-AI/1762652580.020413 | 1762652580.0204139 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | avemio/GRAG-NEMO-12B-ORPO-HESSIAN-AI | avemio/GRAG-NEMO-12B-ORPO-HESSIAN-AI | avemio | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.0}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH", "lower_is_be... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-2.2-rys-78b | cfaafe4c-50a1-4cde-b092-fdbaeea86fb3 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-2.2-rys-78b/1762652579.754511 | 1762652579.754511 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-2.2-rys-78b | MaziyarPanahi/calme-2.2-rys-78b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7986420475449585}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Llama-3-8B-Instruct-v0.8 | c68859dd-6db0-4bdc-a031-92ac7d1d2585 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Llama-3-8B-Instruct-v0.8/1762652579.750486 | 1762652579.750487 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Llama-3-8B-Instruct-v0.8 | MaziyarPanahi/Llama-3-8B-Instruct-v0.8 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7527549125209998}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.1-instruct-78b | 898e5e91-c4c0-4494-baad-37c2bfd1931b | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.1-instruct-78b/1762652579.7584739 | 1762652579.758475 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.1-instruct-78b | MaziyarPanahi/calme-3.1-instruct-78b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8135547015252862}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Llama-3-8B-Instruct-v0.10 | 19143059-07d5-44b2-b599-193147f6196a | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Llama-3-8B-Instruct-v0.10/1762652579.750272 | 1762652579.750272 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Llama-3-8B-Instruct-v0.10 | MaziyarPanahi/Llama-3-8B-Instruct-v0.10 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7667433520835827}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Calme-4x7B-MoE-v0.2 | ca2df1c9-79b2-453b-9cd1-b607e48f5dd7 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Calme-4x7B-MoE-v0.2/1762652579.7498329 | 1762652579.749834 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Calme-4x7B-MoE-v0.2 | MaziyarPanahi/Calme-4x7B-MoE-v0.2 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.429447200095746}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "MixtralForCausalLM", "params_billions": 24.154} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-2.3-rys-78b | 33a06134-e58d-4bc7-8421-c5ae2f0dcd1f | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-2.3-rys-78b/1762652579.7562392 | 1762652579.7562408 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-2.3-rys-78b | MaziyarPanahi/calme-2.3-rys-78b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8065854155862002}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.2-instruct-3b | 83e46bac-5266-4f65-a4dd-76240b297adc | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.2-instruct-3b/1762652579.759095 | 1762652579.7590961 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.2-instruct-3b | MaziyarPanahi/calme-3.2-instruct-3b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5533196363426819}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.3-instruct-3b | 8aa85bd2-eab2-491b-95a3-ac6321cbe298 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.3-instruct-3b/1762652579.759784 | 1762652579.759785 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.3-instruct-3b | MaziyarPanahi/calme-3.3-instruct-3b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6423212631373645}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.2-baguette-3b | e49441f3-99a5-4cdb-bff1-79cc21711bab | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.2-baguette-3b/1762652579.75889 | 1762652579.758891 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.2-baguette-3b | MaziyarPanahi/calme-3.2-baguette-3b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6338282423968404}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.085} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.1-instruct-3b | 67915bce-0b54-4996-90f6-cec6def9bbba | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.1-instruct-3b/1762652579.758249 | 1762652579.75825 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.1-instruct-3b | MaziyarPanahi/calme-3.1-instruct-3b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43359397509718656}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.085} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.1-baguette-3b | 8f0a6518-d153-43ec-b426-02136a2bc367 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.1-baguette-3b/1762652579.7580318 | 1762652579.7580328 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.1-baguette-3b | MaziyarPanahi/calme-3.1-baguette-3b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6234369251364158}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.085} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.2-instruct-78b | 77cc280c-b794-4a9a-addc-e2eb0a1af896 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.2-instruct-78b/1762652579.759298 | 1762652579.759299 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.2-instruct-78b | MaziyarPanahi/calme-3.2-instruct-78b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8062607215521482}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Calme-4x7B-MoE-v0.1 | f4512664-c531-4b13-b76e-e96c2b03febf | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Calme-4x7B-MoE-v0.1/1762652579.7495291 | 1762652579.74953 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Calme-4x7B-MoE-v0.1 | MaziyarPanahi/Calme-4x7B-MoE-v0.1 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4315205875964663}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MixtralForCausalLM", "params_billions": 24.154} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Llama-3-8B-Instruct-v0.9 | 1fb0056b-4f66-404b-89ac-a58185747ce2 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Llama-3-8B-Instruct-v0.9/1762652579.750697 | 1762652579.750697 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Llama-3-8B-Instruct-v0.9 | MaziyarPanahi/Llama-3-8B-Instruct-v0.9 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.763046494412603}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-2.1-rys-78b | 387000a4-7ef5-46c6-9b5e-9bfe7c2cfc18 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-2.1-rys-78b/1762652579.752971 | 1762652579.752971 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-2.1-rys-78b | MaziyarPanahi/calme-2.1-rys-78b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8135547015252862}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Qwen2-7B-Instruct-v0.8 | a65af628-f518-4da7-afc5-7cba4234415b | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Qwen2-7B-Instruct-v0.8/1762652579.751401 | 1762652579.751402 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Qwen2-7B-Instruct-v0.8 | MaziyarPanahi/Qwen2-7B-Instruct-v0.8 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.27747266142723526}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-3.3-baguette-3b | 22cbbb6d-1014-42af-96cf-1636fcb40679 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-3.3-baguette-3b/1762652579.759511 | 1762652579.759511 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-3.3-baguette-3b | MaziyarPanahi/calme-3.3-baguette-3b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6359514975819713}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/calme-2.4-rys-78b | 48433dc8-40ff-4e36-8c6a-ced33bc22e4f | 0.0.1 | hfopenllm_v2/MaziyarPanahi_calme-2.4-rys-78b/1762652579.7570088 | 1762652579.75701 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/calme-2.4-rys-78b | MaziyarPanahi/calme-2.4-rys-78b | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8010899967641414}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 77.965} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Qwen2-7B-Instruct-v0.1 | ce4ee4fe-8a38-467b-b189-b25311c23c4e | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Qwen2-7B-Instruct-v0.1/1762652579.7511811 | 1762652579.751182 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Qwen2-7B-Instruct-v0.1 | MaziyarPanahi/Qwen2-7B-Instruct-v0.1 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.33522498082864577}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | MaziyarPanahi | MaziyarPanahi/Llama-3-70B-Instruct-v0.1 | 1e2759fa-3e87-447b-b0ca-5a4e2e293589 | 0.0.1 | hfopenllm_v2/MaziyarPanahi_Llama-3-70B-Instruct-v0.1/1762652579.750048 | 1762652579.750049 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | MaziyarPanahi/Llama-3-70B-Instruct-v0.1 | MaziyarPanahi/Llama-3-70B-Instruct-v0.1 | MaziyarPanahi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47143800671108216}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 70.554} |
HF Open LLM v2 | PJMixers-Dev | PJMixers-Dev/LLaMa-3.1-Instruct-Interleaved-Zeroed-13B | fb66b283-bfd6-4437-95b7-d74a0d8d2814 | 0.0.1 | hfopenllm_v2/PJMixers-Dev_LLaMa-3.1-Instruct-Interleaved-Zeroed-13B/1762652579.809847 | 1762652579.809848 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PJMixers-Dev/LLaMa-3.1-Instruct-Interleaved-Zeroed-13B | PJMixers-Dev/LLaMa-3.1-Instruct-Interleaved-Zeroed-13B | PJMixers-Dev | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7871015572015585}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 13.047} |
HF Open LLM v2 | PJMixers-Dev | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | 1d91cdce-0bdb-4567-9296-6225db3aa0bc | 0.0.1 | hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B/1762652579.8105159 | 1762652579.810517 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B | PJMixers-Dev | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.693054428915278}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | PJMixers-Dev | PJMixers-Dev/LLaMa-3.2-Instruct-JankMixBread-v0.1-3B | 56f36430-4bb1-425d-ac4b-30d85237667c | 0.0.1 | hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMixBread-v0.1-3B/1762652579.8111491 | 1762652579.81115 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PJMixers-Dev/LLaMa-3.2-Instruct-JankMixBread-v0.1-3B | PJMixers-Dev/LLaMa-3.2-Instruct-JankMixBread-v0.1-3B | PJMixers-Dev | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5040858256093831}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | PJMixers-Dev | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B | d1875dfd-05ab-4a49-8c7f-02cddf35a695 | 0.0.1 | hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B/1762652579.810729 | 1762652579.81073 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-3B | PJMixers-Dev | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6291573026237051}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | PJMixers-Dev | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B | 62b12d95-1da2-407c-8552-8c5e951c5c85 | 0.0.1 | hfopenllm_v2/PJMixers-Dev_LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B/1762652579.8109388 | 1762652579.8109398 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B | PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.2-SFT-HailMary-v0.1-KTO-3B | PJMixers-Dev | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6503898544750152}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | PJMixers-Dev | PJMixers-Dev/L3.2-Instruct-Thinking-v0.1-1B | 3c942d2f-0b53-498e-ab05-71d5075cb974 | 0.0.1 | hfopenllm_v2/PJMixers-Dev_L3.2-Instruct-Thinking-v0.1-1B/1762652579.8095942 | 1762652579.8095949 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PJMixers-Dev/L3.2-Instruct-Thinking-v0.1-1B | PJMixers-Dev/L3.2-Instruct-Thinking-v0.1-1B | PJMixers-Dev | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46276989498973836}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.236} |
HF Open LLM v2 | AicoresSecurity | AicoresSecurity/Cybernet-Sec-3B-R1-V1 | b613ecbe-7b2b-4b03-ab2c-163f9988a8fc | 0.0.1 | hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V1/1762652579.478252 | 1762652579.4782531 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | AicoresSecurity/Cybernet-Sec-3B-R1-V1 | AicoresSecurity/Cybernet-Sec-3B-R1-V1 | AicoresSecurity | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6145693426774292}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | AicoresSecurity | AicoresSecurity/Cybernet-Sec-3B-R1-V0 | 38f169f0-e939-4b12-8f78-b2a27fb90de0 | 0.0.1 | hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V0/1762652579.4777558 | 1762652579.477757 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | AicoresSecurity/Cybernet-Sec-3B-R1-V0 | AicoresSecurity/Cybernet-Sec-3B-R1-V0 | AicoresSecurity | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6358018945287394}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | AicoresSecurity | AicoresSecurity/Cybernet-Sec-3B-R1-V1.1 | e8c63728-a1f5-432f-bf9f-204b0f4041aa | 0.0.1 | hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V1.1/1762652579.478466 | 1762652579.478467 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | AicoresSecurity/Cybernet-Sec-3B-R1-V1.1 | AicoresSecurity/Cybernet-Sec-3B-R1-V1.1 | AicoresSecurity | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6730209178313542}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | AicoresSecurity | AicoresSecurity/Cybernet-Sec-3B-R1-V0-Coder | 48732edf-8baf-438e-8a5c-763eee6c0c18 | 0.0.1 | hfopenllm_v2/AicoresSecurity_Cybernet-Sec-3B-R1-V0-Coder/1762652579.478028 | 1762652579.478029 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | AicoresSecurity/Cybernet-Sec-3B-R1-V0-Coder | AicoresSecurity/Cybernet-Sec-3B-R1-V0-Coder | AicoresSecurity | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7097656440466851}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | tiiuae | tiiuae/falcon-40b | cfdece82-631e-48b7-8232-91a8d9ccf65c | 0.0.1 | hfopenllm_v2/tiiuae_falcon-40b/1762652580.568969 | 1762652580.56897 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/falcon-40b | tiiuae/falcon-40b | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24964538535530173}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 40.0} |
HF Open LLM v2 | tiiuae | tiiuae/falcon-11B | 705a1ff4-2e40-4827-af54-099870fac588 | 0.0.1 | hfopenllm_v2/tiiuae_falcon-11B/1762652580.568774 | 1762652580.568774 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/falcon-11B | tiiuae/falcon-11B | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3261324397044287}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 11.103} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-3B-Base | 1b0d1ae7-322b-46d2-bc33-160f578499b1 | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-3B-Base/1762652580.5675461 | 1762652580.5675468 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-3B-Base | tiiuae/Falcon3-3B-Base | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2764985793250797}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.228} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-10B-Instruct | 741838df-e2a3-4c54-84d3-fe491444071b | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-10B-Instruct/1762652580.566902 | 1762652580.566903 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-10B-Instruct | tiiuae/Falcon3-10B-Instruct | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7816560060639104}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-7B-Instruct | ed988bd0-76b0-4ab6-9c9e-5a5e0aefb936 | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-7B-Instruct/1762652580.568164 | 1762652580.568164 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-7B-Instruct | tiiuae/Falcon3-7B-Instruct | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7612479332615238}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-1B-Base | 1e11a625-87e1-49d0-94a6-8f9ec1f75fc3 | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-1B-Base/1762652580.567122 | 1762652580.567122 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-1B-Base | tiiuae/Falcon3-1B-Base | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24280132271262472}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.669} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-1B-Instruct | a060e2b0-d1ae-48b7-b8f9-c51fadc3e152 | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-1B-Instruct/1762652580.567335 | 1762652580.567335 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-1B-Instruct | tiiuae/Falcon3-1B-Instruct | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5556678501930433}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.669} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-10B-Base | 4e1ce0d3-f454-480b-a4f7-7aa827eaaf1a | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-10B-Base/1762652580.566659 | 1762652580.566659 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-10B-Base | tiiuae/Falcon3-10B-Base | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3647754624396601}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306} |
HF Open LLM v2 | tiiuae | tiiuae/falcon-40b-instruct | 1d6f8802-e9aa-471c-8fbc-1cd807357ab5 | 0.0.1 | hfopenllm_v2/tiiuae_falcon-40b-instruct/1762652580.569173 | 1762652580.569173 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/falcon-40b-instruct | tiiuae/falcon-40b-instruct | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24544874266945038}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 40.0} |
HF Open LLM v2 | tiiuae | tiiuae/falcon-7b | 0e9837cb-4dda-4058-a89e-4127b5980eed | 0.0.1 | hfopenllm_v2/tiiuae_falcon-7b/1762652580.5693781 | 1762652580.569379 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/falcon-7b | tiiuae/falcon-7b | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.182051401392749}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 7.0} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-Mamba-7B-Base | 766e6e63-5779-49cd-9e8c-2bc475c1356a | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-Mamba-7B-Base/1762652580.568367 | 1762652580.5683682 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-Mamba-7B-Base | tiiuae/Falcon3-Mamba-7B-Base | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28911288713945665}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "FalconMambaForCausalLM", "params_billions": 7.273} |
HF Open LLM v2 | tiiuae | tiiuae/falcon-mamba-7b | 9878c419-fff8-402a-a315-70864e5ae60c | 0.0.1 | hfopenllm_v2/tiiuae_falcon-mamba-7b/1762652580.569833 | 1762652580.569834 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/falcon-mamba-7b | tiiuae/falcon-mamba-7b | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3335760227307987}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "FalconMambaForCausalLM", "params_billions": 7.0} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-3B-Instruct | 7aa3aa0e-3b5e-4c0c-a697-2e87859c44f2 | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-3B-Instruct/1762652580.567748 | 1762652580.567749 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-3B-Instruct | tiiuae/Falcon3-3B-Instruct | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6976755010040027}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.228} |
HF Open LLM v2 | tiiuae | tiiuae/falcon-7b-instruct | 2b84722f-58fc-421d-ae1a-9e21ac0b4080 | 0.0.1 | hfopenllm_v2/tiiuae_falcon-7b-instruct/1762652580.5696268 | 1762652580.5696268 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/falcon-7b-instruct | tiiuae/falcon-7b-instruct | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19688869976107837}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "FalconForCausalLM", "params_billions": 7.0} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-7B-Base | 2420519c-81f1-43b3-9b76-af141d2574f4 | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-7B-Base/1762652580.56796 | 1762652580.567961 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-7B-Base | tiiuae/Falcon3-7B-Base | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.34159474638403875}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | tiiuae | tiiuae/Falcon3-Mamba-7B-Instruct | 69491efc-0287-4288-bdf0-bcc57c53b94e | 0.0.1 | hfopenllm_v2/tiiuae_Falcon3-Mamba-7B-Instruct/1762652580.5685718 | 1762652580.5685718 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | tiiuae/Falcon3-Mamba-7B-Instruct | tiiuae/Falcon3-Mamba-7B-Instruct | tiiuae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7165099713205406}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "FalconMambaForCausalLM", "params_billions": 7.273} |
HF Open LLM v2 | alpindale | alpindale/WizardLM-2-8x22B | c2899c4e-5bc9-4b0b-8938-b9848b86fe37 | 0.0.1 | hfopenllm_v2/alpindale_WizardLM-2-8x22B/1762652580.009551 | 1762652580.0095518 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | alpindale/WizardLM-2-8x22B | alpindale/WizardLM-2-8x22B | alpindale | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5272166739805937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MixtralForCausalLM", "params_billions": 140.621} |
HF Open LLM v2 | alpindale | alpindale/magnum-72b-v1 | 186687f8-ed25-44c9-b634-36db1c734844 | 0.0.1 | hfopenllm_v2/alpindale_magnum-72b-v1/1762652580.0098088 | 1762652580.00981 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | alpindale/magnum-72b-v1 | alpindale/magnum-72b-v1 | alpindale | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7606484128778308}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706} |
HF Open LLM v2 | Amaorynho | Amaorynho/BBAI_375 | ad4b6e40-883c-47c5-ba33-6c112c2c6b09 | 0.0.1 | hfopenllm_v2/Amaorynho_BBAI_375/1762652579.480799 | 1762652579.480799 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Amaorynho/BBAI_375 | Amaorynho/BBAI_375 | Amaorynho | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14670518668244703}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.09} |
HF Open LLM v2 | Amaorynho | Amaorynho/BBAI270V4 | 183313de-d526-42a9-a35d-a4e71466e546 | 0.0.1 | hfopenllm_v2/Amaorynho_BBAI270V4/1762652579.4803882 | 1762652579.4803882 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Amaorynho/BBAI270V4 | Amaorynho/BBAI270V4 | Amaorynho | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1990374428737971}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | Amaorynho | Amaorynho/BBAIIFEV1 | 7c0342a3-5bd4-47b0-b238-d5dcb0f6236e | 0.0.1 | hfopenllm_v2/Amaorynho_BBAIIFEV1/1762652579.480599 | 1762652579.4806 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Amaorynho/BBAIIFEV1 | Amaorynho/BBAIIFEV1 | Amaorynho | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8047369867507104}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | Amaorynho | Amaorynho/BBAI2006 | ef37c096-a089-4d3e-9fad-c0f959a18bb3 | 0.0.1 | hfopenllm_v2/Amaorynho_BBAI2006/1762652579.480136 | 1762652579.4801369 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Amaorynho/BBAI2006 | Amaorynho/BBAI2006 | Amaorynho | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14670518668244703}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.09} |
HF Open LLM v2 | NikolaSigmoid | NikolaSigmoid/AceMath-1.5B-Instruct-1epoch | 0cf3db2f-9b23-4602-ac92-265bafd36410 | 0.0.1 | hfopenllm_v2/NikolaSigmoid_AceMath-1.5B-Instruct-1epoch/1762652579.783191 | 1762652579.7831922 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | NikolaSigmoid/AceMath-1.5B-Instruct-1epoch | NikolaSigmoid/AceMath-1.5B-Instruct-1epoch | NikolaSigmoid | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2848918646967823}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.791} |
HF Open LLM v2 | NikolaSigmoid | NikolaSigmoid/acemath-200 | 4414a96e-0664-4531-9c0f-3eb4a062fbe2 | 0.0.1 | hfopenllm_v2/NikolaSigmoid_acemath-200/1762652579.783974 | 1762652579.783974 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | NikolaSigmoid/acemath-200 | NikolaSigmoid/acemath-200 | NikolaSigmoid | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2848918646967823}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.791} |
HF Open LLM v2 | NikolaSigmoid | NikolaSigmoid/AceMath-1.5B-Instruct-dolphin-r1-200 | 93f56942-30d8-4a0f-af8d-901fb264436c | 0.0.1 | hfopenllm_v2/NikolaSigmoid_AceMath-1.5B-Instruct-dolphin-r1-200/1762652579.783446 | 1762652579.783447 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | NikolaSigmoid/AceMath-1.5B-Instruct-dolphin-r1-200 | NikolaSigmoid/AceMath-1.5B-Instruct-dolphin-r1-200 | NikolaSigmoid | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.18080249294095221}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.928} |
HF Open LLM v2 | Alepach | Alepach/notHumpback-M1-v2 | 27c6c36d-6bd5-439b-bdc8-1bd0f8f4c9ea | 0.0.1 | hfopenllm_v2/Alepach_notHumpback-M1-v2/1762652579.4791439 | 1762652579.479145 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Alepach/notHumpback-M1-v2 | Alepach/notHumpback-M1-v2 | Alepach | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2277135777514772}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | Alepach | Alepach/notHumpback-M1 | 030f17b0-036f-4021-90da-6c1d38da659d | 0.0.1 | hfopenllm_v2/Alepach_notHumpback-M1/1762652579.478936 | 1762652579.4789371 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Alepach/notHumpback-M1 | Alepach/notHumpback-M1 | Alepach | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2206944241279804}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | Alepach | Alepach/notHumpback-M0 | 1a4477f7-c414-41ab-bbcb-593f4a86031a | 0.0.1 | hfopenllm_v2/Alepach_notHumpback-M0/1762652579.4786859 | 1762652579.478687 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Alepach/notHumpback-M0 | Alepach/notHumpback-M0 | Alepach | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23500755772461512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | win10 | win10/miscii-14b-1M-0128 | c19f2ddd-7710-4844-9f1f-c0cd1c7e3e41 | 0.0.1 | hfopenllm_v2/win10_miscii-14b-1M-0128/1762652580.5956988 | 1762652580.5957 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | win10/miscii-14b-1M-0128 | win10/miscii-14b-1M-0128 | win10 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4180818007331658}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | win10 | win10/Llama-3.2-3B-Instruct-24-9-29 | bf253a63-4685-4e51-8a0d-5209306926c8 | 0.0.1 | hfopenllm_v2/win10_Llama-3.2-3B-Instruct-24-9-29/1762652580.594629 | 1762652580.59463 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | win10/Llama-3.2-3B-Instruct-24-9-29 | win10/Llama-3.2-3B-Instruct-24-9-29 | win10 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7332211864519476}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | win10 | win10/Breeze-13B-32k-Instruct-v1_0 | bc990db1-c6d9-4113-9946-466bfd5cf9cc | 0.0.1 | hfopenllm_v2/win10_Breeze-13B-32k-Instruct-v1_0/1762652580.5941818 | 1762652580.594183 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | win10/Breeze-13B-32k-Instruct-v1_0 | win10/Breeze-13B-32k-Instruct-v1_0 | win10 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.35843118481185476}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.726} |
HF Open LLM v2 | win10 | win10/llama3-13.45b-Instruct | 3c9eb291-6171-4d40-aa5f-58d39738fdcb | 0.0.1 | hfopenllm_v2/win10_llama3-13.45b-Instruct/1762652580.595499 | 1762652580.5955 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | win10/llama3-13.45b-Instruct | win10/llama3-13.45b-Instruct | win10 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4144348107465968}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 13.265} |
HF Open LLM v2 | win10 | win10/Qwen2.5-2B-Instruct | 143dc973-1063-45d6-9747-9f24a9ae5657 | 0.0.1 | hfopenllm_v2/win10_Qwen2.5-2B-Instruct/1762652580.5952861 | 1762652580.595287 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | win10/Qwen2.5-2B-Instruct | win10/Qwen2.5-2B-Instruct | win10 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22728914834860392}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 2.9} |
HF Open LLM v2 | win10 | win10/ArliAI-RPMax-v1.3-merge-13.3B | 16777b0f-3063-45eb-be07-294d13f975ac | 0.0.1 | hfopenllm_v2/win10_ArliAI-RPMax-v1.3-merge-13.3B/1762652580.593927 | 1762652580.5939279 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | win10/ArliAI-RPMax-v1.3-merge-13.3B | win10/ArliAI-RPMax-v1.3-merge-13.3B | win10 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3038260703821416}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 13.265} |
HF Open LLM v2 | bigscience | bigscience/bloom-3b | 88f90805-7410-4ec1-ad19-8e8a146f1ba3 | 0.0.1 | hfopenllm_v2/bigscience_bloom-3b/1762652580.034177 | 1762652580.034179 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | bigscience/bloom-3b | bigscience/bloom-3b | bigscience | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1270961050013963}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 3.003} |
HF Open LLM v2 | bigscience | bigscience/bloom-1b1 | 284ba4fb-cae4-46ac-a5dd-a36fb145da55 | 0.0.1 | hfopenllm_v2/bigscience_bloom-1b1/1762652580.033589 | 1762652580.033589 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | bigscience/bloom-1b1 | bigscience/bloom-1b1 | bigscience | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13733781920858879}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 1.065} |
HF Open LLM v2 | bigscience | bigscience/bloom-560m | 82454b92-cca1-4ac8-a620-e1a8487a5b8e | 0.0.1 | hfopenllm_v2/bigscience_bloom-560m/1762652580.034546 | 1762652580.034548 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | bigscience/bloom-560m | bigscience/bloom-560m | bigscience | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.06202431769926019}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 0.559} |
HF Open LLM v2 | bigscience | bigscience/bloom-1b7 | 8adb8bb9-d057-45df-827a-cd8f014b4ff6 | 0.0.1 | hfopenllm_v2/bigscience_bloom-1b7/1762652580.033839 | 1762652580.033839 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | bigscience/bloom-1b7 | bigscience/bloom-1b7 | bigscience | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.10438968603305895}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "BloomForCausalLM", "params_billions": 1.722} |
HF Open LLM v2 | bigscience | bigscience/bloom-7b1 | d5fe1452-b6ee-4f1d-9eca-713b49a6a941 | 0.0.1 | hfopenllm_v2/bigscience_bloom-7b1/1762652580.0348449 | 1762652580.034846 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | bigscience/bloom-7b1 | bigscience/bloom-7b1 | bigscience | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13221696210499254}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "BloomForCausalLM", "params_billions": 7.069} |
HF Open LLM v2 | yuvraj17 | yuvraj17/Llama3-8B-SuperNova-Spectrum-Hermes-DPO | d22c83a1-9c1c-43df-b033-c6cb75cb389d | 0.0.1 | hfopenllm_v2/yuvraj17_Llama3-8B-SuperNova-Spectrum-Hermes-DPO/1762652580.611586 | 1762652580.611586 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | yuvraj17/Llama3-8B-SuperNova-Spectrum-Hermes-DPO | yuvraj17/Llama3-8B-SuperNova-Spectrum-Hermes-DPO | yuvraj17 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4690897928607206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | hon9kon9ize | hon9kon9ize/CantoneseLLMChat-v1.0-7B | cccf983e-e1b8-4f0f-b147-abccdea65548 | 0.0.1 | hfopenllm_v2/hon9kon9ize_CantoneseLLMChat-v1.0-7B/1762652580.191013 | 1762652580.191013 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hon9kon9ize/CantoneseLLMChat-v1.0-7B | hon9kon9ize/CantoneseLLMChat-v1.0-7B | hon9kon9ize | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44548353923146145}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | hon9kon9ize | hon9kon9ize/CantoneseLLMChat-v0.5 | 6e87be06-ca0e-48a4-ae28-4a5794600117 | 0.0.1 | hfopenllm_v2/hon9kon9ize_CantoneseLLMChat-v0.5/1762652580.190754 | 1762652580.1907551 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hon9kon9ize/CantoneseLLMChat-v0.5 | hon9kon9ize/CantoneseLLMChat-v0.5 | hon9kon9ize | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3230849701015528}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 6.069} |
HF Open LLM v2 | Khetterman | Khetterman/DarkAtom-12B-v3 | 64802b86-879e-4072-b5ad-aab17d7251f0 | 0.0.1 | hfopenllm_v2/Khetterman_DarkAtom-12B-v3/1762652579.6987362 | 1762652579.698737 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Khetterman/DarkAtom-12B-v3 | Khetterman/DarkAtom-12B-v3 | Khetterman | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6173419859306639}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | Khetterman | Khetterman/Kosmos-8B-v1 | 936cbaa1-e55b-46b8-9610-a5a8faaf4434 | 0.0.1 | hfopenllm_v2/Khetterman_Kosmos-8B-v1/1762652579.6990001 | 1762652579.699001 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Khetterman/Kosmos-8B-v1 | Khetterman/Kosmos-8B-v1 | Khetterman | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41291107594515886}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | RezVortex | RezVortex/JAJUKA-WEWILLNEVERFORGETYOU-3B | 76f26fef-fa87-4cf5-a317-ea4b743e7432 | 0.0.1 | hfopenllm_v2/RezVortex_JAJUKA-WEWILLNEVERFORGETYOU-3B/1762652579.853197 | 1762652579.853197 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | RezVortex/JAJUKA-WEWILLNEVERFORGETYOU-3B | RezVortex/JAJUKA-WEWILLNEVERFORGETYOU-3B | RezVortex | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6858103166265509}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | RezVortex | RezVortex/Jajuka-3b | a41d111c-dd5d-4f77-b52d-9a2dc9f31e50 | 0.0.1 | hfopenllm_v2/RezVortex_Jajuka-3b/1762652579.85344 | 1762652579.853441 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | RezVortex/Jajuka-3b | RezVortex/Jajuka-3b | RezVortex | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6925047762159957}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | formulae | formulae/mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp | 936751f5-4483-4986-9a8c-cb002feb8858 | 0.0.1 | hfopenllm_v2/formulae_mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp/1762652580.1578538 | 1762652580.157855 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp | formulae/mita-elite-sce-gen1.1-v1-7b-2-26-2025-exp | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.16139288199754429}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-elite-v1.1-gen2-7b-2-25-2025 | 106c33d2-84fb-4ea3-b2d3-78981834fdb0 | 0.0.1 | hfopenllm_v2/formulae_mita-elite-v1.1-gen2-7b-2-25-2025/1762652580.158336 | 1762652580.158336 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-elite-v1.1-gen2-7b-2-25-2025 | formulae/mita-elite-v1.1-gen2-7b-2-25-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14108454456397912}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-v1.2-7b-2-24-2025 | a07149d4-66e5-4a0d-b4ae-b696027e821c | 0.0.1 | hfopenllm_v2/formulae_mita-v1.2-7b-2-24-2025/1762652580.160727 | 1762652580.160728 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-v1.2-7b-2-24-2025 | formulae/mita-v1.2-7b-2-24-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.256415200556745}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-elite-v1.2-7b-2-26-2025 | 761560dc-3a0b-481f-8ec2-4d1ea97cfa6f | 0.0.1 | hfopenllm_v2/formulae_mita-elite-v1.2-7b-2-26-2025/1762652580.158752 | 1762652580.158756 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-elite-v1.2-7b-2-26-2025 | formulae/mita-elite-v1.2-7b-2-26-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14800396281865452}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-v1.1-7b-2-24-2025 | 332cbdd8-96b7-40d5-87c6-3610dcbcdc54 | 0.0.1 | hfopenllm_v2/formulae_mita-v1.1-7b-2-24-2025/1762652580.1604211 | 1762652580.1604218 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-v1.1-7b-2-24-2025 | formulae/mita-v1.1-7b-2-24-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.34122018466557624}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-gen3-v1.2-7b-2-26-2025 | a28f8779-d2df-4371-b946-472b335f3ca3 | 0.0.1 | hfopenllm_v2/formulae_mita-gen3-v1.2-7b-2-26-2025/1762652580.15945 | 1762652580.1594508 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-gen3-v1.2-7b-2-26-2025 | formulae/mita-gen3-v1.2-7b-2-26-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2043577707150361}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-gen3-7b-2-26-2025 | 0aa40e02-762d-4a80-932f-f967057c4f50 | 0.0.1 | hfopenllm_v2/formulae_mita-gen3-7b-2-26-2025/1762652580.159164 | 1762652580.159165 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-gen3-7b-2-26-2025 | formulae/mita-gen3-7b-2-26-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1964144026737944}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-math-v2.3-2-25-2025 | fa005333-c7b5-4494-a8cb-4edb1f7d00b9 | 0.0.1 | hfopenllm_v2/formulae_mita-math-v2.3-2-25-2025/1762652580.159737 | 1762652580.159738 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-math-v2.3-2-25-2025 | formulae/mita-math-v2.3-2-25-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13733781920858879}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-v1-7b | 9c629542-6fd0-4cd1-90c7-7f1e95a7a25e | 0.0.1 | hfopenllm_v2/formulae_mita-v1-7b/1762652580.160087 | 1762652580.160088 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-v1-7b | formulae/mita-v1-7b | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19723888172271792}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | formulae | formulae/mita-elite-v1.1-7b-2-25-2025 | 7352f47c-8b57-477f-8190-b08b5b23dfb5 | 0.0.1 | hfopenllm_v2/formulae_mita-elite-v1.1-7b-2-25-2025/1762652580.158112 | 1762652580.158113 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | formulae/mita-elite-v1.1-7b-2-25-2025 | formulae/mita-elite-v1.1-7b-2-25-2025 | formulae | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1249728498162653}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | sometimesanotion | sometimesanotion/lamarck-14b-prose-model_stock | 3191b3a3-761a-42b4-bd31-b8dc22a4c722 | 0.0.1 | hfopenllm_v2/sometimesanotion_lamarck-14b-prose-model_stock/1762652580.5312169 | 1762652580.5312169 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sometimesanotion/lamarck-14b-prose-model_stock | sometimesanotion/lamarck-14b-prose-model_stock | sometimesanotion | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4276486389446668}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | sometimesanotion | sometimesanotion/Lamarck-14B-v0.6-002-model_stock | bd904778-1ad9-48fe-a12e-4b62ce46bd0b | 0.0.1 | hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.6-002-model_stock/1762652580.520087 | 1762652580.520087 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sometimesanotion/Lamarck-14B-v0.6-002-model_stock | sometimesanotion/Lamarck-14B-v0.6-002-model_stock | sometimesanotion | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.669224324791553}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.0} |
HF Open LLM v2 | sometimesanotion | sometimesanotion/Lamarck-14B-v0.7-rc1 | 5919f71f-8d7b-4cce-a7ce-01680c08acf2 | 0.0.1 | hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.7-rc1/1762652580.520714 | 1762652580.520715 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sometimesanotion/Lamarck-14B-v0.7-rc1 | sometimesanotion/Lamarck-14B-v0.7-rc1 | sometimesanotion | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7305482785675341}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | sometimesanotion | sometimesanotion/Lamarck-14B-v0.3 | 6103d107-0eb8-4b0e-8947-d5c7e7cb62f6 | 0.0.1 | hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.3/1762652580.519407 | 1762652580.5194082 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sometimesanotion/Lamarck-14B-v0.3 | sometimesanotion/Lamarck-14B-v0.3 | sometimesanotion | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5031616111916382}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | sometimesanotion | sometimesanotion/Lamarck-14B-v0.1-experimental | aa2b9fb3-77ca-4a48-b3dd-77879220a6b8 | 0.0.1 | hfopenllm_v2/sometimesanotion_Lamarck-14B-v0.1-experimental/1762652580.519198 | 1762652580.519199 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sometimesanotion/Lamarck-14B-v0.1-experimental | sometimesanotion/Lamarck-14B-v0.1-experimental | sometimesanotion | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5353850006870658}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.