Minibase commited on
Commit
3c72e49
·
verified ·
1 Parent(s): 1cec6b3

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +3 -190
config.json CHANGED
@@ -1,190 +1,3 @@
1
- {
2
- "_name_or_path": "Detoxify-Medium",
3
- "architectures": ["LlamaForCausalLM"],
4
- "bos_token_id": 1,
5
- "eos_token_id": 2,
6
- "hidden_act": "silu",
7
- "hidden_size": 1024,
8
- "initializer_range": 0.02,
9
- "intermediate_size": 2816,
10
- "max_position_embeddings": 8192,
11
- "model_type": "llama",
12
- "num_attention_heads": 16,
13
- "num_hidden_layers": 20,
14
- "num_key_value_heads": 16,
15
- "pretraining_tp": 1,
16
- "rms_norm_eps": 1e-05,
17
- "rope_scaling": null,
18
- "rope_theta": 10000.0,
19
- "tie_word_embeddings": false,
20
- "torch_dtype": "float16",
21
- "transformers_version": "4.36.0",
22
- "use_cache": true,
23
- "vocab_size": 49152,
24
-
25
- "quantization_config": {
26
- "quant_method": "gguf",
27
- "quantization_type": "Q8_0",
28
- "quantization_bits": 8,
29
- "quantization_version": "2",
30
- "quantization_fidelity": "high",
31
- "compression_ratio": 0.5,
32
- "quantization_format": "GGUF",
33
- "quantization_description": "8-bit quantization with optimized weights for high fidelity"
34
- },
35
-
36
- "file_size": 387911872,
37
- "file_size_mb": 369,
38
- "context_length": 4096,
39
- "max_seq_length": 4096,
40
- "model_family": "llama",
41
- "model_version": "1.0.0",
42
- "model_size_category": "medium",
43
-
44
- "hardware_requirements": {
45
- "minimum_ram": "12GB",
46
- "recommended_ram": "24GB",
47
- "minimum_vram": "0GB",
48
- "recommended_vram": "0GB",
49
- "cpu_compatibility": "Any modern CPU (Intel i5+, AMD Ryzen 5+)",
50
- "gpu_compatibility": {
51
- "nvidia": "Any CUDA-compatible GPU (GTX 1060+, RTX series)",
52
- "amd": "Any ROCm-compatible GPU (RX 570+, Radeon series)",
53
- "apple_silicon": "M1/M2/M3/M4 chips with Metal (recommended)",
54
- "intel": "Intel Arc GPUs with oneAPI"
55
- },
56
- "os_compatibility": ["linux", "macos", "windows"],
57
- "recommended_hardware": "Apple Silicon M2/M3/M4, Intel i7/i9, or NVIDIA RTX 30-series",
58
- "memory_usage_estimate": "~6-8GB during inference",
59
- "storage_requirements": "400MB free space"
60
- },
61
-
62
- "inference_config": {
63
- "max_tokens": 256,
64
- "temperature": 0.7,
65
- "top_p": 0.9,
66
- "top_k": 40,
67
- "repetition_penalty": 1.1,
68
- "n_threads": 8,
69
- "n_gpu_layers": 99,
70
- "n_batch": 512,
71
- "n_ctx": 4096,
72
- "recommended_batch_size": 1,
73
- "streaming_support": true
74
- },
75
-
76
- "task": "text-detoxification",
77
- "task_description": "Transform toxic or inappropriate text into clean, non-toxic versions while preserving the original meaning and intent",
78
- "task_type": "text2text-generation",
79
- "input_format": "Instruction: {instruction}\\n\\nInput: {input}\\n\\nResponse:",
80
- "output_format": "Clean, non-toxic text that preserves the original meaning",
81
- "primary_use_case": "Content moderation and text detoxification",
82
- "supported_languages": ["en"],
83
-
84
- "performance_metrics": {
85
- "toxicity_reduction": 0.178,
86
- "semantic_preservation": 0.561,
87
- "fluency": 0.929,
88
- "latency_ms": 160.2,
89
- "throughput_requests_per_second": 6.2,
90
- "benchmark_dataset": "ParaDetox",
91
- "benchmark_samples": 1011,
92
- "evaluation_date": "2025-09-18"
93
- },
94
-
95
- "model_capabilities": {
96
- "max_input_length": 4096,
97
- "supports_streaming": true,
98
- "supports_batch_processing": false,
99
- "instruction_following": true,
100
- "few_shot_learning": true,
101
- "multilingual_support": false,
102
- "domain_adaptation": "general_web_content"
103
- },
104
-
105
- "training_details": {
106
- "base_model": "Custom Llama architecture",
107
- "training_objective": "Instruction-following for detoxification",
108
- "fine_tuning_dataset": "Curated toxic-neutral parallel pairs",
109
- "optimization": "Quantized for edge deployment",
110
- "training_framework": "Custom implementation",
111
- "parameter_count": "~150M estimated"
112
- },
113
-
114
- "license": "apache-2.0",
115
- "tags": [
116
- "text-generation",
117
- "text-detoxification",
118
- "toxicity-reduction",
119
- "content-moderation",
120
- "llama",
121
- "gguf",
122
- "quantized",
123
- "cpu-inference",
124
- "apple-silicon",
125
- "minibase",
126
- "medium-model",
127
- "4096-context",
128
- "q8_0"
129
- ],
130
-
131
- "model_creator": "Minibase AI",
132
- "model_url": "https://minibase.ai",
133
- "discord_url": "https://discord.com/invite/BrJn4D2Guh",
134
- "repository_url": "https://github.com/minibase-ai/detoxify-medium",
135
- "paper_url": "https://minibase.ai/research/detoxify-medium",
136
-
137
- "generation_config": {
138
- "_from_model_config": true,
139
- "bos_token_id": 1,
140
- "eos_token_id": 2,
141
- "pad_token_id": null,
142
- "do_sample": true,
143
- "max_new_tokens": 256,
144
- "repetition_penalty": 1.1,
145
- "temperature": 0.7,
146
- "top_k": 40,
147
- "top_p": 0.9,
148
- "transformers_version": "4.36.0"
149
- },
150
-
151
- "tokenizer_config": {
152
- "add_bos_token": true,
153
- "add_eos_token": false,
154
- "bos_token": "<s>",
155
- "eos_token": "</s>",
156
- "unk_token": "<unk>",
157
- "pad_token": null,
158
- "model_max_length": 4096,
159
- "chat_template": "{% for message in messages %}{{ message.content }}{% endfor %}",
160
- "clean_up_tokenization_spaces": false,
161
- "legacy": true
162
- },
163
-
164
- "chat_template": "{% for message in messages %}{{ message.content }}{% endfor %}",
165
- "instruction_template": "Instruction: {instruction}\\n\\nInput: {input}\\n\\nResponse:",
166
-
167
- "safety_and_ethics": {
168
- "intended_use": "Text detoxification and content moderation",
169
- "potential_misuse": "Should not be used for generating toxic content",
170
- "content_policy": "Designed to reduce toxicity in text",
171
- "bias_considerations": "May reflect biases present in training data",
172
- "limitation_awareness": "Not suitable for all languages or contexts"
173
- },
174
-
175
- "technical_support": {
176
- "framework": "llama.cpp",
177
- "runtime_compatibility": ["llama.cpp", "ctransformers", "llama-cpp-python"],
178
- "server_endpoint": "http://127.0.0.1:8000",
179
- "api_format": "OpenAI-compatible",
180
- "streaming_support": true
181
- },
182
-
183
- "version_info": {
184
- "model_version": "1.0.0",
185
- "quantization_version": "Q8_0",
186
- "creation_date": "2025-09-18",
187
- "last_updated": "2025-09-18",
188
- "compatibility_version": "llama.cpp v0.2.0+"
189
- }
190
- }
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d452c14cc5da273518a7957a15bd1db5d44e3d96670f200cf5846a17301676d
3
+ size 5861