_id stringlengths 24 24 | id stringlengths 14 79 | author stringlengths 3 21 | gated stringclasses 3
values | inference stringclasses 9
values | lastModified stringlengths 24 24 | likes int64 0 4.13k | private bool 1
class | sha stringlengths 40 40 | config stringlengths 23 5.92k ⌀ | downloads int64 1.8k 3.13M | tags sequencelengths 5 62 | pipeline_tag stringclasses 6
values | library_name stringclasses 4
values | createdAt stringlengths 24 24 | modelId stringlengths 14 79 | siblings listlengths 4 285 | model_type stringclasses 13
values | base_model stringlengths 13 43 | base_model_downloads float64 98 28.9k | base_model_pipeline stringclasses 1
value | base_model_children_count float64 98 28.9k | adapter float64 0 377 | merge float64 0 94 | finetune float64 0 263 | quantized float64 0 182 | __index_level_0__ int64 86k 992k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6698d8a0653e4babe21e1e7d | meta-llama/Llama-3.1-8B-Instruct | meta-llama | manual | loading | 2024-09-25T17:00:57.000Z | 2,649 | false | 0e9e39f249a16976918f6564b8830bc894c89659 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag... | 3,134,749 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2204.05149",
"base_model:meta-llama/Llama-3.1-8B",
"base_model:finetune:meta-llama/Llama-3.1-8B",
"lic... | text-generation | transformers | 2024-07-18T08:56:00.000Z | meta-llama/Llama-3.1-8B-Instruct | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
... | llama | meta-llama/Llama-3.1-8B | 254 | text-generation | 254 | 377 | 22 | 241 | 182 | 760,607 |
667928d58601cd537f63c846 | google/gemma-2-9b-it | google | manual | warm | 2024-08-27T19:41:49.000Z | 468 | false | 11c9b309abf73637e4b6f9a3fa1e92e615547819 | {"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0... | 964,871 | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:2009.03300",
"arxiv:1905.07830",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1905.10044",
"arxiv:1907.10641",
"arxiv:1811.00937",
"arxiv:1809.02789",
"arxiv:1911.01547",
"arxiv:1705.03551",
"arxiv:2... | text-generation | transformers | 2024-06-24T08:05:41.000Z | google/gemma-2-9b-it | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "mo... | gemma2 | google/gemma-2-9b | 105 | text-generation | 105 | 25 | 6 | 33 | 69 | 718,125 |
653a4cbd66081db6fc588b21 | HuggingFaceH4/zephyr-7b-beta | HuggingFaceH4 | False | warm | 2024-09-23T13:33:11.000Z | 1,578 | false | 20e1a5880bb00a7571542fe3fe6cb2dcb4816eee | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['conte... | 749,121 | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"arxiv:2305.18290",
"arxiv:2310.16944",
"base_model:mistralai/Mistral-7B-v0.1",
... | text-generation | transformers | 2023-10-26T11:25:49.000Z | HuggingFaceH4/zephyr-7b-beta | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 274 | 76 | 120 | 20 | 357,190 |
657607b0a90ae2daaef65917 | mistralai/Mixtral-8x7B-Instruct-v0.1 | mistralai | auto | warm | 2024-08-19T13:18:42.000Z | 4,130 | false | 41bd4c9e7e4fb318ca40e721131d4933966c2cc1 | {"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages ... | 578,188 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:finetune:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoint... | text-generation | transformers | 2023-12-10T18:47:12.000Z | mistralai/Mixtral-8x7B-Instruct-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "consolidated.00.pt"
},
{
"rfilename": "consolidated.01.pt"
},
{
"rfilename": "consolidated.02.pt"
},
{
"rfilename": "consolidated.03.pt"
},
{
"... | mixtral | mistralai/Mixtral-8x7B-v0.1 | 182 | text-generation | 182 | 110 | 48 | 33 | 20 | 413,206 |
664dc170474f2283fa5c8659 | mistralai/Mistral-7B-Instruct-v0.3 | mistralai | auto | warm | 2024-08-21T12:18:25.000Z | 998 | false | e0bc86c23ce5aae1db576c8cca6f06f1f73af2db | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0][\"role\"] == \"system\" %}\n {%- set system_message = messages[0][\"content\"] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = mes... | 564,853 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:mistralai/Mistral-7B-v0.3",
"base_model:finetune:mistralai/Mistral-7B-v0.3",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-22T09:57:04.000Z | mistralai/Mistral-7B-Instruct-v0.3 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "consolidated.safetensors"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-0000... | mistral | mistralai/Mistral-7B-v0.3 | 338 | text-generation | 338 | 181 | 10 | 51 | 87 | 652,648 |
65143cd8e31c0e2e3df713e5 | mistralai/Mistral-7B-Instruct-v0.1 | mistralai | auto | warm | 2024-08-22T11:33:25.000Z | 1,506 | false | 2dcff66eac0c01dc50e4c41eea959968232187fe | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages ... | 562,402 | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"conversational",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"en... | text-generation | transformers | 2023-09-27T14:31:52.000Z | mistralai/Mistral-7B-Instruct-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "mo... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 340 | 41 | 139 | 14 | 326,808 |
665ee74789b4fd787a568664 | Qwen/Qwen2-7B-Instruct | Qwen | False | cold | 2024-08-21T10:29:04.000Z | 573 | false | f2826a00ceef68f0f2b946d945ecc0477ce4450c | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] +... | 410,195 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-04T10:07:03.000Z | Qwen/Qwen2-7B-Instruct | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"r... | qwen2 | Qwen/Qwen2-7B | 266 | text-generation | 266 | 49 | 3 | 42 | 69 | 680,671 |
669650bb11dbbf600cf4dcf0 | google/gemma-2-2b-it | google | manual | warm | 2024-08-27T19:41:44.000Z | 570 | false | 299a8560bedf22ed1c72a8a11e7dce4a7f9f51f8 | {"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0... | 358,990 | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:2009.03300",
"arxiv:1905.07830",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1905.10044",
"arxiv:1907.10641",
"arxiv:1811.00937",
"arxiv:1809.02789",
"arxiv:1911.01547",
"arxiv:1705.03551",
"arxiv:2... | text-generation | transformers | 2024-07-16T10:51:39.000Z | google/gemma-2-2b-it | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "mo... | gemma2 | google/gemma-2-2b | 256 | text-generation | 256 | 120 | 3 | 74 | 77 | 755,923 |
664e5306762e0826aea132d0 | thesven/Mistral-7B-Instruct-v0.3-GPTQ | thesven | False | cold | 2024-09-11T17:17:42.000Z | 0 | false | 2d24a7f41b3d93fba3aef0cde15a77f09e99437c | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversa... | 345,189 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:mistralai/Mistral-7B-Instruct-v0.3",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.3",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",... | text-generation | transformers | 2024-05-22T20:18:14.000Z | thesven/Mistral-7B-Instruct-v0.3-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
... | mistral | mistralai/Mistral-7B-Instruct-v0.3 | 320 | text-generation | 320 | 0 | 0 | 0 | 0 | 653,744 |
6515ff5b7f18cec973af0828 | TheBloke/Mistral-7B-Instruct-v0.1-GPTQ | TheBloke | False | explicit-opt-out | 2023-09-29T20:48:48.000Z | 75 | false | 6ae1e4ae2cfbaf107c705ed722ec243b4f88014d | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversa... | 296,813 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"conversational",
"base_model:mistralai/Mistral-7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
... | text-generation | transformers | 2023-09-28T22:34:03.000Z | TheBloke/Mistral-7B-Instruct-v0.1-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},... | mistral | mistralai/Mistral-7B-Instruct-v0.1 | 500 | text-generation | 500 | 191 | 0 | 92 | 0 | 328,345 |
65cac0d27faf059c56a5821f | google/gemma-7b-it | google | manual | cold | 2024-08-14T08:36:20.000Z | 1,128 | false | 9c5798d27f588501ce1e108079d2a19e4c3a2353 | {"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 %... | 273,413 | [
"transformers",
"safetensors",
"gguf",
"gemma",
"text-generation",
"conversational",
"arxiv:2312.11805",
"arxiv:2009.03300",
"arxiv:1905.07830",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1905.10044",
"arxiv:1907.10641",
"arxiv:1811.00937",
"arxiv:1809.02789",
"arxiv:1911.01547",
... | text-generation | transformers | 2024-02-13T01:07:30.000Z | google/gemma-7b-it | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "gemma-7b-it.gguf"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-000... | gemma | google/gemma-7b | 9,278 | text-generation | 9,278 | 74 | 0 | 16 | 10 | 490,051 |
665d877e957df09a0751515e | Qwen/Qwen2-0.5B-Instruct | Qwen | False | cold | 2024-08-21T10:23:36.000Z | 155 | false | c540970f9e29518b1d8f06ab8b24cba66ad77b6d | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] +... | 270,072 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"base_model:Qwen/Qwen2-0.5B",
"base_model:finetune:Qwen/Qwen2-0.5B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-03T09:06:06.000Z | Qwen/Qwen2-0.5B-Instruct | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "tok... | qwen2 | Qwen/Qwen2-0.5B | 120 | text-generation | 120 | 162 | 4 | 35 | 47 | 678,240 |
669fe28bb3b73c95ecd1b6cd | SanctumAI/Meta-Llama-3.1-8B-Instruct-GGUF | SanctumAI | False | cold | 2024-09-11T17:15:23.000Z | 10 | false | 1cbd69377fb06c0a4719187fc31fdc7d4020ec3a | {"model_type": "llama"} | 269,192 | [
"transformers",
"gguf",
"llama",
"facebook",
"meta",
"pytorch",
"llama-3",
"text-generation",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:quantized:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3.1",
"text-genera... | text-generation | transformers | 2024-07-23T17:04:11.000Z | SanctumAI/Meta-Llama-3.1-8B-Instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "meta-llama-3.1-8b-instruct.Q2_K.gguf"
},
{
"rfilename": "meta-llama-3.1-8b-instruct.Q3_K_L.gguf"
},
{
"rfilename": "meta-llama-3.1-8b-instruct.Q3_K_M.gguf"
}... | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 772,940 |
65778ac662d3ac1817cc9201 | TheBloke/Mistral-7B-Instruct-v0.2-GGUF | TheBloke | False | explicit-opt-out | 2023-12-11T22:23:10.000Z | 383 | false | 3a6fbf4a41a1d52e415a4958cde6856d34b2db93 | {"model_type": "mistral"} | 212,551 | [
"transformers",
"gguf",
"mistral",
"finetuned",
"text-generation",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-12-11T22:18:46.000Z | TheBloke/Mistral-7B-Instruct-v0.2-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q2_K.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_L.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_M.gguf"
},
{
... | mistral | mistralai/Mistral-7B-Instruct-v0.2 | 1,321 | text-generation | 1,321 | 0 | 0 | 0 | 0 | 414,766 |
664e8ee4437ce2c3146cb273 | neuralmagic/Mistral-7B-Instruct-v0.3-GPTQ-4bit | neuralmagic | False | cold | 2024-06-10T20:59:32.000Z | 11 | false | ed07c8f1d2c87648508d9e153199d9b761cad5a8 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversa... | 186,770 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:mistralai/Mistral-7B-Instruct-v0.3",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.3",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compat... | text-generation | transformers | 2024-05-23T00:33:40.000Z | neuralmagic/Mistral-7B-Instruct-v0.3-GPTQ-4bit | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
... | mistral | mistralai/Mistral-7B-Instruct-v0.3 | 320 | text-generation | 320 | 0 | 0 | 0 | 0 | 653,995 |
666154bbf57dbfbdc3696bd2 | Qwen/Qwen2-7B-Instruct-AWQ | Qwen | False | cold | 2024-08-21T10:30:36.000Z | 19 | false | f7621c98023fa293e743c692981a3c8f5b564b86 | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im... | 184,675 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"base_model:Qwen/Qwen2-7B-Instruct",
"base_model:quantized:Qwen/Qwen2-7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatibl... | text-generation | transformers | 2024-06-06T06:18:35.000Z | Qwen/Qwen2-7B-Instruct-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"r... | qwen2 | Qwen/Qwen2-7B-Instruct | 161 | text-generation | 161 | 0 | 0 | 0 | 0 | 685,154 |
65148287463c4fd767f76cb6 | TheBloke/Mistral-7B-Instruct-v0.1-AWQ | TheBloke | False | explicit-opt-out | 2023-11-09T18:17:58.000Z | 35 | false | b2f7c152209c12057c3a0d77b2c01a1def7d594f | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"quant_method": "awq", "bits": 4}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": true}} | 178,358 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"base_model:mistralai/Mistral-7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"awq",
"region:us"
] | text-generation | transformers | 2023-09-27T19:29:11.000Z | TheBloke/Mistral-7B-Instruct-v0.1-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quant_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
... | mistral | mistralai/Mistral-7B-Instruct-v0.1 | 500 | text-generation | 500 | 0 | 0 | 0 | 0 | 327,065 |
669fbfba1631541594c89063 | casperhansen/mistral-nemo-instruct-2407-awq | casperhansen | False | cold | 2024-09-27T07:14:03.000Z | 5 | false | c83b6438e13051ad1c0f5683635705ee83bb8772 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages... | 165,811 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:mistralai/Mistral-Nemo-Instruct-2407",
"base_model:quantized:mistralai/Mistral-Nemo-Instruct-2407",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"awq",
"region:... | text-generation | transformers | 2024-07-23T14:35:38.000Z | casperhansen/mistral-nemo-instruct-2407-awq | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "mo... | mistral | mistralai/Mistral-Nemo-Instruct-2407 | 101 | text-generation | 101 | 0 | 0 | 0 | 0 | 772,614 |
66231225c8920ec35160357e | aaditya/Llama3-OpenBioLLM-8B | aaditya | False | cold | 2024-04-28T02:36:57.000Z | 146 | false | 000c725dc3a680e35260b2c213163387581c974f | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}} | 162,094 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"llama-3",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"en",
"arxiv:2305.18290",
"arxiv:2303.13375",
"arxiv:2212.13138",
"arxiv:2305.09617",
"arxiv:2402.07023",
"base_model:meta-llama... | text-generation | transformers | 2024-04-20T00:53:57.000Z | aaditya/Llama3-OpenBioLLM-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "py... | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 30 | 8 | 4 | 594,742 |
6515ffbc3e5a12e0ceaefd75 | TheBloke/Mistral-7B-v0.1-GPTQ | TheBloke | False | explicit-opt-out | 2023-09-29T20:49:41.000Z | 35 | false | 81de15eeac5938bc3b4065dfddf798fe5d215881 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": true}} | 149,047 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"pretrained",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:quantized:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | transformers | 2023-09-28T22:35:40.000Z | TheBloke/Mistral-7B-v0.1-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 25 | 0 | 5 | 1 | 328,348 |
66968801480e8b04bb69248c | SanctumAI/gemma-2-9b-it-GGUF | SanctumAI | False | cold | 2024-09-11T17:17:18.000Z | 2 | false | 5e3f967050720200fe63b2f8a62e3f6ac5c096ea | {"model_type": "gemma2"} | 130,857 | [
"transformers",
"gguf",
"gemma2",
"conversational",
"text-generation",
"base_model:google/gemma-2-9b-it",
"base_model:quantized:google/gemma-2-9b-it",
"license:gemma",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-16T14:47:29.000Z | SanctumAI/gemma-2-9b-it-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "gemma-2-9b-it.Q2_K.gguf"
},
{
"rfilename": "gemma-2-9b-it.Q3_K_L.gguf"
},
{
"rfilename": "gemma-2-9b-it.Q3_K_M.gguf"
},
{
"rfilename": "gemma-2-9b-it.Q... | gemma2 | google/gemma-2-9b-it | 131 | text-generation | 131 | 0 | 0 | 0 | 0 | 756,278 |
64b6ce072edcb4a1aa820cbe | TheBloke/Llama-2-7B-Chat-GPTQ | TheBloke | False | explicit-opt-out | 2023-09-27T12:44:48.000Z | 258 | false | d5ad9310836dd91b6ac6133e2e47f47394386cea | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "cont... | 110,199 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-2",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-7b-chat-hf",
"license:llama2",
"autotrain_compatible",
"text-generation-infe... | text-generation | transformers | 2023-07-18T17:38:15.000Z | TheBloke/Llama-2-7B-Chat-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safet... | llama | meta-llama/Llama-2-7b-chat-hf | 1,328 | text-generation | 1,328 | 13 | 0 | 3 | 0 | 247,838 |
66303852b8d7220d89548ffc | NousResearch/Hermes-2-Pro-Llama-3-8B | NousResearch | False | cold | 2024-09-14T16:29:41.000Z | 400 | false | f798274b30e7d2d4797c369edcc0cc7473b6e6f2 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": [{"name": "default", "template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_g... | 97,818 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"axolotl",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:Nou... | text-generation | transformers | 2024-04-30T00:16:18.000Z | NousResearch/Hermes-2-Pro-Llama-3-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "dpo-adapter/adapter_config.json"
},
{
"rfilename": "dpo-adapter/adapter_model.safetensors"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename":... | llama | NousResearch/Meta-Llama-3-8B | 101 | text-generation | 101 | 2 | 32 | 73 | 35 | 612,280 |
662a2d8b084d6b7ab42d8a15 | MLP-KTLim/llama-3-Korean-Bllossom-8B | MLP-KTLim | False | cold | 2024-08-12T04:01:02.000Z | 258 | false | 10acb1aa4f341f2d3c899d78c520b0822a909b95 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|s... | 85,510 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"ko",
"arxiv:2403.10882",
"arxiv:2403.11399",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",... | text-generation | transformers | 2024-04-25T10:16:43.000Z | MLP-KTLim/llama-3-Korean-Bllossom-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "mo... | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 13 | 13 | 22 | 28 | 604,596 |
64f60811b8cc49b414fe5cdf | TheBloke/Llama-2-7B-Chat-GGUF | TheBloke | False | explicit-opt-out | 2023-10-14T21:36:33.000Z | 426 | false | 191239b3e26b2882fb562ffccdd1cf0f65402adb | {"model_type": "llama"} | 79,952 | [
"transformers",
"gguf",
"llama",
"facebook",
"meta",
"pytorch",
"llama-2",
"text-generation",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-7b-chat-hf",
"license:llama2",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-09-04T16:38:41.000Z | TheBloke/Llama-2-7B-Chat-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "llama-2-7b-chat.Q2_K.gguf"
},
{
"rfilename": "llam... | llama | meta-llama/Llama-2-7b-chat-hf | 1,328 | text-generation | 1,328 | 0 | 0 | 0 | 0 | 303,198 |
657759d1da3f14fa5ddeba81 | TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ | TheBloke | False | explicit-opt-out | 2023-12-14T14:30:44.000Z | 130 | false | 0f81ba4680ccd2bce163334b93305d40b9e27b09 | {"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false, "chat_template": "{{ bos_token }}{% for message in mes... | 70,053 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-infe... | text-generation | transformers | 2023-12-11T18:49:53.000Z | TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},... | mixtral | mistralai/Mixtral-8x7B-Instruct-v0.1 | 166 | text-generation | 166 | 0 | 0 | 0 | 0 | 414,590 |
664f44644f17b67fdf917d9d | SanctumAI/Mistral-7B-Instruct-v0.3-GGUF | SanctumAI | False | cold | 2024-09-15T11:33:21.000Z | 3 | false | ab77e5f8958a1e3b612b805f89f4695e8451950a | {"model_type": "mistral"} | 67,526 | [
"transformers",
"gguf",
"mistral",
"text-generation",
"base_model:mistralai/Mistral-7B-Instruct-v0.3",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.3",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-23T13:28:04.000Z | SanctumAI/Mistral-7B-Instruct-v0.3-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mistral-7b-instruct-v0.3.Q2_K.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.3.Q3_K_L.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.3.Q3_K_M.gguf"
},
{
... | mistral | mistralai/Mistral-7B-Instruct-v0.3 | 320 | text-generation | 320 | 0 | 0 | 0 | 0 | 654,981 |
6591dd37a02954c9825f00f1 | TheBloke/TinyLlama-1.1B-Chat-v1.0-GPTQ | TheBloke | False | explicit-opt-out | 2023-12-31T21:33:52.000Z | 12 | false | 9d4580af0f21bccafd762dcc50d0c7bac6273584 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['r... | 67,033 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"en",
"dataset:cerebras/SlimPajama-627B",
"dataset:bigcode/starcoderdata",
"dataset:OpenAssistant/oasst_top1_2023-08-25",
"base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"base_model:quantized:TinyLlama/TinyLlama-1.1B-C... | text-generation | transformers | 2023-12-31T21:29:27.000Z | TheBloke/TinyLlama-1.1B-Chat-v1.0-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
... | llama | TinyLlama/TinyLlama-1.1B-Chat-v1.0 | 687 | text-generation | 687 | 0 | 0 | 1 | 0 | 436,937 |
66cbcec8e0cdc14fc0b45840 | smeby/Qwen-Qwen1.5-7B-1724632776 | smeby | False | pipeline-not-detected | 2024-08-26T00:39:52.000Z | 0 | false | 45923cca577d6ac53d6b9c212c8c6e07ebd65d97 | {"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{... | 49,744 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"base_model:adapter:Qwen/Qwen1.5-7B",
"region:us"
] | null | peft | 2024-08-26T00:39:36.000Z | smeby/Qwen-Qwen1.5-7B-1724632776 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "checkpoint-31/README.md"
},
{
"rfilename": "checkpoint-31/a... | null | Qwen/Qwen1.5-7B | 6,515 | text-generation | 6,515 | 0 | 0 | 0 | 0 | 860,315 |
65778ac63ceeb2f078e6c864 | TheBloke/Mistral-7B-Instruct-v0.2-GPTQ | TheBloke | False | explicit-opt-out | 2023-12-11T22:46:53.000Z | 48 | false | 7532d6bc89ef9300fb39d2d94ed4414ec534b72a | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false, "chat_template": "{{ bos_token }}{% for message in mes... | 47,926 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"conversational",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",... | text-generation | transformers | 2023-12-11T22:18:46.000Z | TheBloke/Mistral-7B-Instruct-v0.2-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},... | mistral | mistralai/Mistral-7B-Instruct-v0.2 | 1,321 | text-generation | 1,321 | 269 | 0 | 2 | 1 | 414,764 |
65778ac658d7a2cc892266d6 | TheBloke/Mistral-7B-Instruct-v0.2-AWQ | TheBloke | False | explicit-opt-out | 2023-12-11T22:36:20.000Z | 41 | false | f970a2bb89d5c2f9d217dc337f39e24625d6462a | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false, "chat_template": "{{ bos_token }}{% for message in mess... | 46,303 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"finetuned",
"conversational",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",... | text-generation | transformers | 2023-12-11T22:18:46.000Z | TheBloke/Mistral-7B-Instruct-v0.2-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quant_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
... | mistral | mistralai/Mistral-7B-Instruct-v0.2 | 1,321 | text-generation | 1,321 | 15 | 0 | 0 | 0 | 414,765 |
659a5e051b4a26041af869ac | openchat/openchat-3.5-0106 | openchat | False | not-popular-enough | 2024-05-18T18:14:51.000Z | 342 | false | ff058fda49726ecf4ea53dc1635f917cdb8ba36b | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{{ 'GPT4 Correct ' + message['role'].title() + ': ' + message['content'] + '<|end_of_turn|>'}}{% endfor %}{% if add_generation_prompt %}{{ 'GPT4 Corre... | 45,756 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"openchat",
"C-RLFT",
"conversational",
"arxiv:2309.11235",
"arxiv:2303.08774",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generati... | text-generation | transformers | 2024-01-07T08:17:09.000Z | openchat/openchat-3.5-0106 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 16 | 28 | 30 | 22 | 443,802 |
66616ddf3fdeeb338135dbef | Qwen/Qwen2-0.5B-Instruct-GGUF | Qwen | False | library-not-detected | 2024-08-21T10:29:53.000Z | 54 | false | 198f08841147e5196a6a69bd0053690fb1fd3857 | null | 45,743 | [
"gguf",
"instruct",
"chat",
"text-generation",
"en",
"base_model:Qwen/Qwen2-0.5B-Instruct",
"base_model:quantized:Qwen/Qwen2-0.5B-Instruct",
"license:apache-2.0",
"region:us"
] | text-generation | null | 2024-06-06T08:05:51.000Z | Qwen/Qwen2-0.5B-Instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "qwen2-0_5b-instruct-fp16.gguf"
},
{
"rfilename": "qwen2-0_5b-instruct-q2_k.gguf"
},
{
"rfilename": "qwen2-0_5b-instruct-q3_k_m.gguf"
},
{
"rfilename": "qwe... | null | Qwen/Qwen2-0.5B-Instruct | 178 | text-generation | 178 | 0 | 0 | 0 | 0 | 685,351 |
653ec25781277ed9683d246f | teknium/OpenHermes-2.5-Mistral-7B | teknium | False | not-popular-enough | 2024-02-19T17:53:06.000Z | 803 | false | 24c0bea14d53e6f67f1fbe2eca5bfe7cae389b33 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "<|im_end|>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_s... | 45,623 | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7... | text-generation | transformers | 2023-10-29T20:36:39.000Z | teknium/OpenHermes-2.5-Mistral-7B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 96 | 37 | 71 | 30 | 360,799 |
65775021c991ca09567c1237 | TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF | TheBloke | False | explicit-opt-out | 2023-12-14T14:30:43.000Z | 596 | false | fa1d3835c5d45a3a74c0b68805fcdc133dba2b6a | {"model_type": "mixtral"} | 45,482 | [
"transformers",
"gguf",
"mixtral",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] | null | transformers | 2023-12-11T18:08:33.000Z | TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mixtral-8x7b-instruct-v0.1.Q2_K.gguf"
},
{
"rfilename": "mixtral-8x7b-instruct-v0.1.Q3_K_M.gguf"
},
{
"rfilename": "mixtral-8x7b-instruct-v0.1.Q4_0.gguf"
},
... | mixtral | mistralai/Mixtral-8x7B-Instruct-v0.1 | 166 | text-generation | 166 | 0 | 0 | 0 | 0 | 414,555 |
66a5de99a28bc058db44d78f | NousResearch/Hermes-3-Llama-3.1-8B | NousResearch | False | warm | 2024-09-08T07:39:55.000Z | 195 | false | 896ea440e5a9e6070e3d8a2774daf2b481ab425b | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": [{"name": "default", "template": "{{bos_token}}{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|... | 42,774 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Llama-3",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"axolotl",
"roleplaying",
"chat",
"conversational",
"en",
"arxiv:2408.11857",
"base_model:meta-llam... | text-generation | transformers | 2024-07-28T06:00:57.000Z | NousResearch/Hermes-3-Llama-3.1-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "mo... | llama | meta-llama/Llama-3.1-8B | 254 | text-generation | 254 | 1 | 16 | 10 | 33 | 782,842 |
663238fcd5677e328794061c | DeepMount00/Llama-3-8b-Ita | DeepMount00 | False | not-popular-enough | 2024-08-13T11:16:20.000Z | 23 | false | 9e5753beda3704492ae340e0321ad516a58b9ffd | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|e... | 40,242 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"it",
"en",
"dataset:DeepMount00/llm_ita_ultra",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation... | text-generation | transformers | 2024-05-01T12:43:40.000Z | DeepMount00/Llama-3-8b-Ita | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "mo... | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 43 | 0 | 2 | 614,878 |
66d8ab8cebc4415df3ff3836 | modularai/llama-3.1 | modularai | False | explicit-opt-out | 2024-09-09T16:49:02.000Z | 0 | false | 966694508430d1177f6d585de779250e7a34bc3a | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag... | 29,726 | [
"gguf",
"llama",
"facebook",
"meta",
"pytorch",
"llama-3",
"text-generation",
"conversational",
"en",
"arxiv:2204.05149",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:quantized:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3",
"region:us"
] | text-generation | null | 2024-09-04T18:48:44.000Z | modularai/llama-3.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "llama-3.1-8b-instruct-bf16.gguf"
},
{
"rfilename": "llama-3.1-8b-instruct-f32.gguf"
},
{
"rfilename": "llama-3.1-8b-instruc... | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 898,568 |
6591d4d754f88261730df832 | TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF | TheBloke | False | explicit-opt-out | 2023-12-31T21:29:33.000Z | 110 | false | 52e7645ba7c309695bec7ac98f4f005b139cf465 | {"model_type": "tinyllama"} | 28,570 | [
"transformers",
"gguf",
"tinyllama",
"en",
"dataset:cerebras/SlimPajama-627B",
"dataset:bigcode/starcoderdata",
"dataset:OpenAssistant/oasst_top1_2023-08-25",
"base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"base_model:quantized:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"license:apache-2.0",
"region... | null | transformers | 2023-12-31T20:53:43.000Z | TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "tinyllama-1.1b-chat-v1.0.Q2_K.gguf"
},
{
"rfilename": "tinyllama-1.1b-chat-v1.0.Q3_K_L.gguf"
},
{
"rfilename": "tinyllama-1.1b-chat-v1.0.Q3_K_M.gguf"
},
{
... | tinyllama | TinyLlama/TinyLlama-1.1B-Chat-v1.0 | 687 | text-generation | 687 | 0 | 0 | 2 | 0 | 436,921 |
657f5104476260623ddf5081 | mlc-ai/Llama-2-7b-chat-hf-q4f32_1-MLC | mlc-ai | False | pipeline-not-detected | 2024-07-11T15:31:37.000Z | 2 | false | bab021ce3e8b5715dc8679df1661c32a5d1e26da | {"tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token... | 28,221 | [
"mlc-llm",
"web-llm",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-7b-chat-hf",
"region:us"
] | null | mlc-llm | 2023-12-17T19:50:28.000Z | mlc-ai/Llama-2-7b-chat-hf-q4f32_1-MLC | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "logs.txt"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"r... | null | meta-llama/Llama-2-7b-chat-hf | 1,328 | text-generation | 1,328 | 0 | 0 | 0 | 0 | 422,014 |
66a2ab4c3c7f0decff7157d6 | lmstudio-community/Phi-3.1-mini-128k-instruct-GGUF | lmstudio-community | False | library-not-detected | 2024-07-25T19:49:22.000Z | 2 | false | 6fb38f94bbd8e88338e0b965d056e0d780b5f8b7 | null | 25,199 | [
"gguf",
"nlp",
"code",
"text-generation",
"en",
"arxiv:2404.14219",
"base_model:microsoft/Phi-3-mini-128k-instruct",
"base_model:quantized:microsoft/Phi-3-mini-128k-instruct",
"license:mit",
"region:us"
] | text-generation | null | 2024-07-25T19:45:16.000Z | lmstudio-community/Phi-3.1-mini-128k-instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_M.gguf"
},
{
... | null | microsoft/Phi-3-mini-128k-instruct | 196 | text-generation | 196 | 0 | 0 | 0 | 0 | 778,040 |
654cce913a321f06c2b5ba05 | alignment-handbook/zephyr-7b-sft-full | alignment-handbook | False | not-popular-enough | 2024-01-10T03:18:43.000Z | 21 | false | 92f9fac4529acacb2c33a35c46917393690c6311 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['conte... | 24,850 | [
"transformers",
"tensorboard",
"safetensors",
"mistral",
"text-generation",
"alignment-handbook",
"generated_from_trainer",
"trl",
"sft",
"conversational",
"dataset:HuggingFaceH4/ultrachat_200k",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"lice... | text-generation | transformers | 2023-11-09T12:20:33.000Z | alignment-handbook/zephyr-7b-sft-full | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 136 | 0 | 263 | 1 | 372,399 |
65146b42c75a3d4c44e41667 | TheBloke/Mistral-7B-Instruct-v0.1-GGUF | TheBloke | False | explicit-opt-out | 2023-12-09T16:09:28.000Z | 508 | false | 731a9fc8f06f5f5e2db8a0cf9d256197eb6e05d1 | {"model_type": "mistral"} | 24,376 | [
"transformers",
"gguf",
"mistral",
"finetuned",
"text-generation",
"base_model:mistralai/Mistral-7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.1",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-09-27T17:49:54.000Z | TheBloke/Mistral-7B-Instruct-v0.1-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mistral-7b-instruct-v0.1.Q2_K.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.1.Q3_K_L.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.1.Q3_K_M.gguf"
},
{
... | mistral | mistralai/Mistral-7B-Instruct-v0.1 | 500 | text-generation | 500 | 0 | 0 | 0 | 0 | 326,999 |
6659b8b93114a8eca3bcb84a | numind/NuExtract-tiny | numind | False | not-popular-enough | 2024-08-23T14:49:20.000Z | 35 | false | c4fca7faff3f8ef4f020fd22cf4480b7f2d82c14 | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + ... | 23,877 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"base_model:Qwen/Qwen1.5-0.5B",
"base_model:finetune:Qwen/Qwen1.5-0.5B",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-31T11:47:05.000Z | numind/NuExtract-tiny | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model.safetensors"
},
{
"rfilen... | qwen2 | Qwen/Qwen1.5-0.5B | 28,947 | text-generation | 28,947 | 0 | 0 | 7 | 3 | 672,306 |
667d887808289e6bddd446c7 | bartowski/gemma-2-9b-it-GGUF | bartowski | False | not-popular-enough | 2024-07-15T18:57:00.000Z | 177 | false | d731033f3dc4018261fd39896e50984d398b4ac5 | null | 22,919 | [
"transformers",
"gguf",
"conversational",
"text-generation",
"base_model:google/gemma-2-9b-it",
"base_model:quantized:google/gemma-2-9b-it",
"license:gemma",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-27T15:42:48.000Z | bartowski/gemma-2-9b-it-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "gemma-2-9b-it-IQ2_M.gguf"
},
{
"rfilename": "gemma-2-9b-it-IQ2_S.gguf"
},
{
"rfilename": "gemma-2-9b-it-IQ2_XS.gguf"
},
{
"rfilename": "gemma-2-9b-it-IQ3_M.gguf"
},
{
"rfilename": "ge... | null | google/gemma-2-9b-it | 131 | text-generation | 131 | 0 | 0 | 0 | 0 | 725,091 |
664e2b118e43ee411f6ca3c4 | MaziyarPanahi/Mistral-7B-Instruct-v0.3-GGUF | MaziyarPanahi | False | explicit-opt-out | 2024-05-22T20:26:06.000Z | 58 | false | ce89f595755a4bf2e2e05d155cc43cb847c78978 | {"model_type": "mistral"} | 22,372 | [
"transformers",
"gguf",
"mistral",
"quantized",
"2-bit",
"3-bit",
"4-bit",
"5-bit",
"6-bit",
"8-bit",
"GGUF",
"safetensors",
"text-generation",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us",
"bas... | text-generation | transformers | 2024-05-22T17:27:45.000Z | MaziyarPanahi/Mistral-7B-Instruct-v0.3-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Mistral-7B-Instruct-v0.3.IQ1_M.gguf"
},
{
"rfilename": "Mistral-7B-Instruct-v0.3.IQ1_S.gguf"
},
{
"rfilename": "Mistral-7B-Instruct-v0.3.IQ2_XS.gguf"
},
{
"rfilename": "Mistral-7B-Instruct-v0.3.IQ3_XS.gguf"
},
{
"rfilenam... | mistral | mistralai/Mistral-7B-Instruct-v0.3 | 320 | text-generation | 320 | 0 | 0 | 0 | 0 | 653,473 |
66aa6a19a9ec1810a37eb785 | bartowski/gemma-2-2b-it-GGUF | bartowski | False | not-popular-enough | 2024-08-05T19:14:36.000Z | 26 | false | 855f67caed130e1befc571b52bd181be2e858883 | null | 22,089 | [
"transformers",
"gguf",
"conversational",
"text-generation",
"base_model:google/gemma-2-2b-it",
"base_model:quantized:google/gemma-2-2b-it",
"license:gemma",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-31T16:45:13.000Z | bartowski/gemma-2-2b-it-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "gemma-2-2b-it-IQ3_M.gguf"
},
{
"rfilename": "gemma-2-2b-it-IQ4_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it-Q3_K_L.gguf"
},
{
"rfilename": "gemma-2-2b-it-Q4_K_M.gguf"
},
{
"rfilename": "... | null | google/gemma-2-2b-it | 265 | text-generation | 265 | 0 | 0 | 0 | 0 | 791,308 |
6523bd962d3ba46ccd378b5a | HuggingFaceH4/zephyr-7b-alpha | HuggingFaceH4 | False | loading | 2023-11-21T17:28:11.000Z | 1,094 | false | 2ce2d025864af849b3e5029e2ec9d568eeda892d | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['conte... | 22,023 | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:stingning/ultrachat",
"dataset:openbmb/UltraFeedback",
"arxiv:2305.18290",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1... | text-generation | transformers | 2023-10-09T08:45:10.000Z | HuggingFaceH4/zephyr-7b-alpha | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "colab-demo.ipynb"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilen... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 42 | 1 | 12 | 8 | 337,729 |
66de024731e772c5ec381060 | silent666/01-ai-Yi-1.5-9B-1725825607 | silent666 | False | pipeline-not-detected | 2024-09-08T20:00:49.000Z | 0 | false | 86e187005fff940a3f68ccd12dd0083deb7a96c0 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 20,833 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-08T20:00:07.000Z | silent666/01-ai-Yi-1.5-9B-1725825607 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.j... | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 916,071 |
66c4f4e7f77e8e56f3806368 | bartowski/Phi-3.5-mini-instruct-GGUF | bartowski | False | not-popular-enough | 2024-09-15T07:35:15.000Z | 36 | false | 6d70da17e749a471ccb62ade694486011a75cda3 | null | 19,761 | [
"transformers",
"gguf",
"nlp",
"code",
"text-generation",
"multilingual",
"base_model:microsoft/Phi-3.5-mini-instruct",
"base_model:quantized:microsoft/Phi-3.5-mini-instruct",
"license:mit",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-08-20T19:56:23.000Z | bartowski/Phi-3.5-mini-instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.5-mini-instruct-IQ2_M.gguf"
},
{
"rfilename": "Phi-3.5-mini-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.5-mini-instruct-IQ3_XS.gguf"
},
{
"rfilename": "Phi-3.5-mini-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.5... | null | microsoft/Phi-3.5-mini-instruct | 147 | text-generation | 147 | 0 | 0 | 0 | 0 | 842,257 |
6696a3096fa71ac401b70de0 | princeton-nlp/gemma-2-9b-it-SimPO | princeton-nlp | False | not-popular-enough | 2024-08-02T22:04:11.000Z | 80 | false | 8c87091f412e3aa6f74f66bd86c57fb81cbc3fde | {"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] | trim + '\n\n' %}{% set messages = messages[1:] %}{% else %}{% set system_message = '' %}{%... | 19,520 | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"alignment-handbook",
"generated_from_trainer",
"conversational",
"dataset:princeton-nlp/gemma2-ultrafeedback-armorm",
"arxiv:2405.14734",
"base_model:google/gemma-2-9b-it",
"base_model:finetune:google/gemma-2-9b-it",
"license:mit",
... | text-generation | transformers | 2024-07-16T16:42:49.000Z | princeton-nlp/gemma-2-9b-it-SimPO | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"... | gemma2 | google/gemma-2-9b-it | 131 | text-generation | 131 | 0 | 17 | 8 | 15 | 756,458 |
6661b70d346ffbe5a3d44175 | Qwen/Qwen2-7B-Instruct-GGUF | Qwen | False | library-not-detected | 2024-08-21T10:28:11.000Z | 163 | false | c3024c6fff0a02d52119ecee024bbb93d4b4b8e4 | null | 19,088 | [
"gguf",
"chat",
"text-generation",
"en",
"base_model:Qwen/Qwen2-7B-Instruct",
"base_model:quantized:Qwen/Qwen2-7B-Instruct",
"license:apache-2.0",
"region:us"
] | text-generation | null | 2024-06-06T13:18:05.000Z | Qwen/Qwen2-7B-Instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "qwen2-7b-instruct-fp16.gguf"
},
{
"rfilename": "qwen2-7b-instruct-q2_k.gguf"
},
{
"rfilename": "qwen2-7b-instruct-q3_k_m.gguf"
},
{
"rfilename": "qwen2-7b-... | null | Qwen/Qwen2-7B-Instruct | 161 | text-generation | 161 | 0 | 0 | 0 | 0 | 685,781 |
64f5fd954d3b1dd311d30e28 | TheBloke/Llama-2-7B-GGUF | TheBloke | False | explicit-opt-out | 2023-10-24T07:32:45.000Z | 173 | false | b4e04e128f421c93a5f1e34ac4d7ca9b0af47b80 | {"model_type": "llama"} | 18,814 | [
"transformers",
"gguf",
"llama",
"facebook",
"meta",
"pytorch",
"llama-2",
"text-generation",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:quantized:meta-llama/Llama-2-7b-hf",
"license:llama2",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-09-04T15:53:57.000Z | TheBloke/Llama-2-7B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "llama-2-7b.Q2_K.gguf"
},
{
"rfilename": "llama-2-7... | llama | meta-llama/Llama-2-7b-hf | 1,595 | text-generation | 1,595 | 0 | 0 | 0 | 0 | 303,156 |
64b6c679d16e945399007406 | TheBloke/Llama-2-7B-GPTQ | TheBloke | False | explicit-opt-out | 2023-09-27T12:44:46.000Z | 80 | false | 60b5c75d47a7d925782e74d16b6686cf0abbd052 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "cont... | 18,449 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-2",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:quantized:meta-llama/Llama-2-7b-hf",
"license:llama2",
"autotrain_compatible",
"text-generation-inference",
... | text-generation | transformers | 2023-07-18T17:06:01.000Z | TheBloke/Llama-2-7B-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safet... | llama | meta-llama/Llama-2-7b-hf | 1,595 | text-generation | 1,595 | 5 | 0 | 1 | 1 | 247,793 |
6631ab448d7c840b14eb5840 | NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF | NousResearch | False | library-not-detected | 2024-05-03T14:21:35.000Z | 154 | false | fc0d3245797976d0effe82a4f6803a5814e6a4fe | null | 18,414 | [
"gguf",
"Llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:NousResearch/Meta-Llama-3-8B",
"base_model:quantized:NousResearch/Meta-Llama-3-8B",
"lice... | null | null | 2024-05-01T02:39:00.000Z | NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-8B-F16.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-8B-Q5_K_M.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-8B-Q6_K.gguf"
},
{
"rfilename": "He... | null | NousResearch/Meta-Llama-3-8B | 101 | text-generation | 101 | 0 | 0 | 0 | 0 | 614,232 |
65780e5e353869cd6ef1d1ca | openchat/openchat-3.5-1210 | openchat | False | not-popular-enough | 2024-05-18T18:10:44.000Z | 276 | false | 801f5459b7577241500785f11c2b026912badd6e | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{{ 'GPT4 Correct ' + message['role'].title() + ': ' + message['content'] + '<|end_of_turn|>'}}{% endfor %}{% if add_generation_prompt %}{{ 'GPT4 Corre... | 18,317 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"openchat",
"C-RLFT",
"conversational",
"dataset:openchat/openchat_sharegpt4_dataset",
"dataset:kaist-ai/Feedback-Collection",
"dataset:imone/OpenOrca_FLAN",
"dataset:LDJnr/Capybara",
"dataset:tiedong/goat",
"dataset:glaiveai/glaiv... | text-generation | transformers | 2023-12-12T07:40:14.000Z | openchat/openchat-3.5-1210 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 8 | 14 | 11 | 5 | 415,184 |
66286775d59deb30bb03684f | maum-ai/Llama-3-MAAL-8B-Instruct-v0.1 | maum-ai | False | not-popular-enough | 2024-04-30T12:25:15.000Z | 31 | false | f6e6495a184ac245c8e3c953ce6acb8f732c52dd | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|e... | 18,172 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"llama-3",
"llama-3-ko",
"conversational",
"en",
"ko",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"autotrain_compatible",
"... | text-generation | transformers | 2024-04-24T01:59:17.000Z | maum-ai/Llama-3-MAAL-8B-Instruct-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "mo... | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 12 | 0 | 0 | 601,965 |
66aa7b285bb7aca6cc409418 | lmstudio-community/gemma-2-2b-it-GGUF | lmstudio-community | False | not-popular-enough | 2024-07-31T18:58:29.000Z | 16 | false | 6aa72da804ad76c5dc862867bfba6256de9172c7 | null | 16,615 | [
"transformers",
"gguf",
"conversational",
"text-generation",
"base_model:google/gemma-2-2b-it",
"base_model:quantized:google/gemma-2-2b-it",
"license:gemma",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-31T17:58:00.000Z | lmstudio-community/gemma-2-2b-it-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "gemma-2-2b-it-IQ3_M.gguf"
},
{
"rfilename": "gemma-2-2b-it-IQ4_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it-Q3_K_L.gguf"
},
{
"rfilename": "gemma-2-2b-it-Q4_K_M.gguf"
},
{
"rfilename": "... | null | google/gemma-2-2b-it | 265 | text-generation | 265 | 0 | 0 | 0 | 0 | 791,420 |
65eeff3ffc59f6e77eafb754 | NousResearch/Hermes-2-Pro-Mistral-7B | NousResearch | False | not-popular-enough | 2024-09-08T08:08:34.000Z | 480 | false | 24dbda51d986efa05201eec7a093ac91fc5519cf | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": [{"name": "default", "template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_... | 16,543 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"Mistral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mist... | text-generation | transformers | 2024-03-11T12:55:27.000Z | NousResearch/Hermes-2-Pro-Mistral-7B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00... | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 2 | 94 | 14 | 15 | 530,374 |
66dd9c3efd9b7f48c89ca228 | amiguel/classItem-FT-llama-3-1-8b-instruct | amiguel | False | not-popular-enough | 2024-09-08T12:47:25.000Z | 0 | false | 82d8630451db886a42309d2088f6c5b1d9d3a3e3 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"load_in_4bit": true, "load_in_8bit": false, "quant_method": "bitsandbytes"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_to... | 16,524 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"conversational",
"en",
"base_model:unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit",
"base_model:quantized:unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit",
"license:apache-2.0",
"auto... | text-generation | transformers | 2024-09-08T12:44:46.000Z | amiguel/classItem-FT-llama-3-1-8b-instruct | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "mo... | llama | unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit | 661 | text-generation | 661 | 0 | 0 | 0 | 0 | 914,454 |
64b6d9d4bb1d8f2832c8be93 | TheBloke/Llama-2-13B-chat-GPTQ | TheBloke | False | explicit-opt-out | 2023-09-27T12:44:48.000Z | 358 | false | ea078917a7e91c896787c73dba935f032ae658e9 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "cont... | 16,056 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-2",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-13b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-13b-chat-hf",
"license:llama2",
"autotrain_compatible",
"text-generation-in... | text-generation | transformers | 2023-07-18T18:28:36.000Z | TheBloke/Llama-2-13B-chat-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safet... | llama | meta-llama/Llama-2-13b-chat-hf | 318 | text-generation | 318 | 5 | 0 | 0 | 0 | 247,882 |
66693891e06cc524750de8f7 | QuantFactory/Qwen2-0.5B-GGUF | QuantFactory | False | library-not-detected | 2024-06-18T06:31:15.000Z | 2 | false | 2a3c6a4d2fa2077c6a3045867ac18b1933fef5be | null | 15,865 | [
"gguf",
"pretrained",
"text-generation",
"en",
"base_model:Qwen/Qwen2-0.5B",
"base_model:quantized:Qwen/Qwen2-0.5B",
"license:apache-2.0",
"region:us"
] | text-generation | null | 2024-06-12T05:56:33.000Z | QuantFactory/Qwen2-0.5B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Qwen2-0.5B.Q2_K.gguf"
},
{
"rfilename": "Qwen2-0.5B.Q3_K_L.gguf"
},
{
"rfilename": "Qwen2-0.5B.Q3_K_M.gguf"
},
{
"rfilename": "Qwen2-0.5B.Q3_K_S.gguf"
},
{
"rfilename": "Qwen2-0.5B.Q4_0.gguf"
},
{
"rfilename": "Qw... | null | Qwen/Qwen2-0.5B | 120 | text-generation | 120 | 0 | 0 | 0 | 0 | 695,554 |
66a4fbbe07dd6d0ab4a07650 | GPT4All-Community/Meta-Llama-3.1-8B-Instruct-128k-GGUF | GPT4All-Community | False | library-not-detected | 2024-08-12T23:51:19.000Z | 8 | false | 350b6d7f3a2224c98b6dc8ebdce0e290b71cae22 | null | 15,300 | [
"gguf",
"text-generation-inference",
"text-generation",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:quantized:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3.1",
"region:us"
] | text-generation | null | 2024-07-27T13:53:02.000Z | GPT4All-Community/Meta-Llama-3.1-8B-Instruct-128k-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Meta-Llama-3.1-8B-Instruct-128k-Q4_0.gguf"
},
{
"rfilename": "Meta-Llama-3.1-8B-Instruct-128k-f16.gguf"
},
{
"rfilename": "README.md"
}
] | null | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 781,715 |
6508c16eabdde5290e606c80 | TheBloke/Llama-2-7B-Chat-AWQ | TheBloke | False | explicit-opt-out | 2023-11-09T18:21:14.000Z | 21 | false | ee91097dc818c742d9b648147891dfadd9638ace | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"quant_method": "awq", "bits": 4}, "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "cont... | 14,453 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-2",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:quantized:meta-llama/Llama-2-7b-chat-hf",
"license:llama2",
"autotrain_compatible",
"text-generation-infe... | text-generation | transformers | 2023-09-18T21:30:22.000Z | TheBloke/Llama-2-7B-Chat-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.s... | llama | meta-llama/Llama-2-7b-chat-hf | 1,328 | text-generation | 1,328 | 0 | 0 | 0 | 0 | 317,895 |
65d445c10e77d4f229c98fdc | NousResearch/Nous-Hermes-2-Mistral-7B-DPO-GGUF | NousResearch | False | library-not-detected | 2024-02-21T01:35:04.000Z | 59 | false | eb85cf06e8663157611e8ee472e61b43f50ee49f | null | 14,221 | [
"gguf",
"Mistral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:quantized:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"region:us"
] | null | null | 2024-02-20T06:25:05.000Z | NousResearch/Nous-Hermes-2-Mistral-7B-DPO-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Nous-Hermes-2-Mistral-7B-DPO.Q2_K.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mistral-7B-DPO.Q3_K_L.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mistral-7B-DPO.Q3_K_M.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mistral-7B-DPO.Q3_K_S.gguf"
},
... | null | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 0 | 0 | 0 | 500,060 |
6627253888596f571aedfb4a | lightblue/suzume-llama-3-8B-multilingual | lightblue | False | not-popular-enough | 2024-06-02T02:14:24.000Z | 103 | false | 0cb15aa9ec685eef494f9a15f65aefcfe3c04c66 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|e... | 13,616 | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"conversational",
"arxiv:2405.12612",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:other",
"autotrain_compatible",
"text-ge... | text-generation | transformers | 2024-04-23T03:04:24.000Z | lightblue/suzume-llama-3-8B-multilingual | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "mo... | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 3 | 7 | 10 | 600,092 |
666283eda5e2fb9d6b7336c5 | gaianet/Qwen2-0.5B-Instruct-GGUF | gaianet | False | not-popular-enough | 2024-06-07T04:09:30.000Z | 1 | false | bf2944f900c00233d7d1251f4584616cbb06883e | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2"} | 13,032 | [
"transformers",
"gguf",
"qwen2",
"text-generation",
"chat",
"en",
"base_model:Qwen/Qwen2-0.5B-Instruct",
"base_model:quantized:Qwen/Qwen2-0.5B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-07T03:52:13.000Z | gaianet/Qwen2-0.5B-Instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Qwen2-0.5B-Instruct-Q2_K.gguf"
},
{
"rfilename": "Qwen2-0.5B-Instruct-Q3_K_L.gguf"
},
{
"rfilename": "Qwen2-0.5B-Instruct-Q3_K_M.gguf"
},
{
"rfilename": "Qwen2-0.5B-Instruct-Q3_K_S.gguf"
},
{
"rfilename": "Qwen2-0.5B-Inst... | qwen2 | Qwen/Qwen2-0.5B-Instruct | 178 | text-generation | 178 | 0 | 0 | 0 | 0 | 686,950 |
66374df5650f76807b257b98 | NousResearch/Hermes-2-Theta-Llama-3-8B | NousResearch | False | not-popular-enough | 2024-09-08T08:11:20.000Z | 192 | false | 57a73110702e7b05ba3f39fef36297454c680725 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": [{"name": "default", "template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_g... | 13,026 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"axolotl",
"merges",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"ba... | text-generation | transformers | 2024-05-05T09:14:29.000Z | NousResearch/Hermes-2-Theta-Llama-3-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "mo... | llama | NousResearch/Hermes-2-Pro-Llama-3-8B | 131 | text-generation | 131 | 0 | 23 | 1 | 9 | 621,829 |
66615487a4c55c16aceb4423 | Qwen/Qwen2-1.5B-Instruct-AWQ | Qwen | False | not-popular-enough | 2024-08-21T10:35:28.000Z | 7 | false | 688a0bab90cef22c0a5fad44dbda2b65a8859147 | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im... | 12,374 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"base_model:Qwen/Qwen2-1.5B-Instruct",
"base_model:quantized:Qwen/Qwen2-1.5B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"... | text-generation | transformers | 2024-06-06T06:17:43.000Z | Qwen/Qwen2-1.5B-Instruct-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "tok... | qwen2 | Qwen/Qwen2-1.5B-Instruct | 495 | text-generation | 495 | 0 | 0 | 0 | 0 | 685,151 |
663e0c516757087d1abb4ab2 | MaziyarPanahi/calme-2.3-phi3-4b | MaziyarPanahi | False | explicit-opt-out | 2024-07-20T13:39:11.000Z | 9 | false | 2f0dfbb0426ae75b0106f260658ca155b99a8df5 | {"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% if not add_... | 12,004 | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"axolotl",
"finetune",
"dpo",
"microsoft",
"phi",
"pytorch",
"phi-3",
"nlp",
"code",
"chatml",
"conversational",
"custom_code",
"en",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:finetune:microsoft/Phi-3-mini... | text-generation | transformers | 2024-05-10T12:00:17.000Z | MaziyarPanahi/calme-2.3-phi3-4b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00... | phi3 | microsoft/Phi-3-mini-4k-instruct | 470 | text-generation | 470 | 0 | 0 | 3 | 3 | 631,158 |
66214d85acbb21038cda695a | QuantFactory/Meta-Llama-3-8B-GGUF | QuantFactory | False | library-not-detected | 2024-04-20T16:20:22.000Z | 107 | false | 1ca85c857dce892b673b988ad0aa83f2cb1bbd19 | null | 11,654 | [
"gguf",
"facebook",
"meta",
"pytorch",
"llama",
"llama-3",
"text-generation",
"en",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:quantized:meta-llama/Meta-Llama-3-8B",
"license:other",
"region:us"
] | text-generation | null | 2024-04-18T16:42:45.000Z | QuantFactory/Meta-Llama-3-8B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "Meta-Llama-3-8B.Q2_K.gguf"
},
{
"rfilename": "Meta-Llama-3-8B.Q3_K_L.gguf"
},
{
"rfilename": "Meta-Llama-3-8B.Q3_K_M.gguf"
},
{
"rfilename": "Meta-Llama-3-8B.Q3_K_S.gguf"
},
{
"rfil... | null | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 0 | 592,376 |
End of preview. Expand in Data Studio
README.md exists but content is empty.
- Downloads last month
- 11