Trouter-Library commited on
Commit
72b94b9
·
verified ·
1 Parent(s): d115a04

Create metadata.json

Browse files
Files changed (1) hide show
  1. metadata.json +168 -0
metadata.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "Helion-V2",
3
+ "model_version": "2.0",
4
+ "model_type": "causal-lm",
5
+ "architecture": "helion",
6
+ "organization": "DeepXR",
7
+ "release_date": "2024-11-15",
8
+ "license": "Apache-2.0",
9
+ "homepage": "https://huggingface.co/DeepXR/Helion-V2",
10
+ "repository": "https://github.com/DeepXR/Helion-V2",
11
+ "documentation": "https://docs.deepxr.ai/helion-v2",
12
+ "parameters": {
13
+ "total": 7200000000,
14
+ "trainable": 7200000000,
15
+ "non_trainable": 0
16
+ },
17
+ "model_size": {
18
+ "fp32": "28.8 GB",
19
+ "fp16": "14.4 GB",
20
+ "bfloat16": "14.4 GB",
21
+ "int8": "7.2 GB",
22
+ "int4": "3.6 GB"
23
+ },
24
+ "context_length": 8192,
25
+ "vocabulary_size": 32768,
26
+ "languages": [
27
+ "en",
28
+ "es",
29
+ "fr",
30
+ "de",
31
+ "it",
32
+ "pt",
33
+ "nl",
34
+ "ru",
35
+ "zh",
36
+ "ja",
37
+ "ko",
38
+ "ar",
39
+ "hi"
40
+ ],
41
+ "primary_language": "en",
42
+ "tasks": [
43
+ "text-generation",
44
+ "conversational",
45
+ "code-generation",
46
+ "question-answering",
47
+ "summarization",
48
+ "translation",
49
+ "creative-writing"
50
+ ],
51
+ "framework": {
52
+ "name": "transformers",
53
+ "version": "4.40.0+",
54
+ "backend": "pytorch",
55
+ "backend_version": "2.1.0+"
56
+ },
57
+ "training": {
58
+ "dataset_tokens": 2500000000000,
59
+ "training_steps": 600000,
60
+ "epochs": 3,
61
+ "batch_size": 4194304,
62
+ "learning_rate": 0.0003,
63
+ "optimizer": "AdamW",
64
+ "hardware": "128x NVIDIA H100 80GB",
65
+ "training_time_days": 21,
66
+ "knowledge_cutoff": "2024-10-31"
67
+ },
68
+ "benchmarks": {
69
+ "mmlu": {
70
+ "score": 64.2,
71
+ "metric": "accuracy",
72
+ "shots": 5
73
+ },
74
+ "humaneval": {
75
+ "score": 48.2,
76
+ "metric": "pass@1",
77
+ "shots": 0
78
+ },
79
+ "hellaswag": {
80
+ "score": 80.5,
81
+ "metric": "accuracy",
82
+ "shots": 10
83
+ },
84
+ "truthfulqa": {
85
+ "score": 52.1,
86
+ "metric": "mc2_accuracy",
87
+ "shots": 0
88
+ },
89
+ "gsm8k": {
90
+ "score": 68.7,
91
+ "metric": "accuracy",
92
+ "shots": 8
93
+ },
94
+ "arc_challenge": {
95
+ "score": 58.3,
96
+ "metric": "accuracy",
97
+ "shots": 25
98
+ },
99
+ "mt_bench": {
100
+ "score": 7.85,
101
+ "metric": "rating",
102
+ "shots": 0
103
+ }
104
+ },
105
+ "safety": {
106
+ "toxigen_score": 0.08,
107
+ "crows_pairs_bias": 54.2,
108
+ "content_filtering": true,
109
+ "pii_detection": true,
110
+ "crisis_detection": true,
111
+ "safety_classifiers": [
112
+ "hate_speech",
113
+ "violence",
114
+ "sexual_content",
115
+ "self_harm",
116
+ "illegal_activity",
117
+ "child_safety"
118
+ ]
119
+ },
120
+ "deployment": {
121
+ "recommended_gpu": [
122
+ "NVIDIA A100",
123
+ "NVIDIA H100",
124
+ "NVIDIA RTX 4090",
125
+ "NVIDIA A6000"
126
+ ],
127
+ "minimum_vram_gb": 16,
128
+ "recommended_vram_gb": 24,
129
+ "supports_quantization": true,
130
+ "quantization_formats": [
131
+ "int8",
132
+ "int4",
133
+ "GPTQ",
134
+ "AWQ",
135
+ "GGUF"
136
+ ],
137
+ "inference_frameworks": [
138
+ "transformers",
139
+ "vllm",
140
+ "llama.cpp",
141
+ "text-generation-inference"
142
+ ]
143
+ },
144
+ "contact": {
145
+ "email": "contact@deepxr.ai",
146
+ "github": "https://github.com/DeepXR/Helion-V2",
147
+ "twitter": "@DeepXR_AI",
148
+ "discord": "https://discord.gg/deepxr"
149
+ },
150
+ "tags": [
151
+ "text-generation",
152
+ "conversational",
153
+ "code",
154
+ "instruction-following",
155
+ "pytorch",
156
+ "causal-lm",
157
+ "llm",
158
+ "reasoning",
159
+ "multilingual",
160
+ "apache-2.0",
161
+ "production-ready",
162
+ "safe-ai"
163
+ ],
164
+ "checksum": {
165
+ "algorithm": "SHA256",
166
+ "model_files": "computed_on_upload"
167
+ }
168
+ }