squ11z1 commited on
Commit
e357168
·
verified ·
1 Parent(s): c351f80

Delete lora

Browse files
lora/README.md DELETED
@@ -1,210 +0,0 @@
1
- ---
2
- base_model: Jackrong/Qwen3.5-9B-DeepSeek-V4-Flash
3
- library_name: peft
4
- pipeline_tag: text-generation
5
- tags:
6
- - base_model:adapter:Jackrong/Qwen3.5-9B-DeepSeek-V4-Flash
7
- - lora
8
- - sft
9
- - transformers
10
- - trl
11
- - unsloth
12
- ---
13
-
14
- # Model Card for Model ID
15
-
16
- <!-- Provide a quick summary of what the model is/does. -->
17
-
18
-
19
-
20
- ## Model Details
21
-
22
- ### Model Description
23
-
24
- <!-- Provide a longer summary of what this model is. -->
25
-
26
-
27
-
28
- - **Developed by:** [More Information Needed]
29
- - **Funded by [optional]:** [More Information Needed]
30
- - **Shared by [optional]:** [More Information Needed]
31
- - **Model type:** [More Information Needed]
32
- - **Language(s) (NLP):** [More Information Needed]
33
- - **License:** [More Information Needed]
34
- - **Finetuned from model [optional]:** [More Information Needed]
35
-
36
- ### Model Sources [optional]
37
-
38
- <!-- Provide the basic links for the model. -->
39
-
40
- - **Repository:** [More Information Needed]
41
- - **Paper [optional]:** [More Information Needed]
42
- - **Demo [optional]:** [More Information Needed]
43
-
44
- ## Uses
45
-
46
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
47
-
48
- ### Direct Use
49
-
50
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
51
-
52
- [More Information Needed]
53
-
54
- ### Downstream Use [optional]
55
-
56
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
57
-
58
- [More Information Needed]
59
-
60
- ### Out-of-Scope Use
61
-
62
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
63
-
64
- [More Information Needed]
65
-
66
- ## Bias, Risks, and Limitations
67
-
68
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
69
-
70
- [More Information Needed]
71
-
72
- ### Recommendations
73
-
74
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
75
-
76
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
77
-
78
- ## How to Get Started with the Model
79
-
80
- Use the code below to get started with the model.
81
-
82
- [More Information Needed]
83
-
84
- ## Training Details
85
-
86
- ### Training Data
87
-
88
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
89
-
90
- [More Information Needed]
91
-
92
- ### Training Procedure
93
-
94
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
95
-
96
- #### Preprocessing [optional]
97
-
98
- [More Information Needed]
99
-
100
-
101
- #### Training Hyperparameters
102
-
103
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
104
-
105
- #### Speeds, Sizes, Times [optional]
106
-
107
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
108
-
109
- [More Information Needed]
110
-
111
- ## Evaluation
112
-
113
- <!-- This section describes the evaluation protocols and provides the results. -->
114
-
115
- ### Testing Data, Factors & Metrics
116
-
117
- #### Testing Data
118
-
119
- <!-- This should link to a Dataset Card if possible. -->
120
-
121
- [More Information Needed]
122
-
123
- #### Factors
124
-
125
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
126
-
127
- [More Information Needed]
128
-
129
- #### Metrics
130
-
131
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
132
-
133
- [More Information Needed]
134
-
135
- ### Results
136
-
137
- [More Information Needed]
138
-
139
- #### Summary
140
-
141
-
142
-
143
- ## Model Examination [optional]
144
-
145
- <!-- Relevant interpretability work for the model goes here -->
146
-
147
- [More Information Needed]
148
-
149
- ## Environmental Impact
150
-
151
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
152
-
153
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
154
-
155
- - **Hardware Type:** [More Information Needed]
156
- - **Hours used:** [More Information Needed]
157
- - **Cloud Provider:** [More Information Needed]
158
- - **Compute Region:** [More Information Needed]
159
- - **Carbon Emitted:** [More Information Needed]
160
-
161
- ## Technical Specifications [optional]
162
-
163
- ### Model Architecture and Objective
164
-
165
- [More Information Needed]
166
-
167
- ### Compute Infrastructure
168
-
169
- [More Information Needed]
170
-
171
- #### Hardware
172
-
173
- [More Information Needed]
174
-
175
- #### Software
176
-
177
- [More Information Needed]
178
-
179
- ## Citation [optional]
180
-
181
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
182
-
183
- **BibTeX:**
184
-
185
- [More Information Needed]
186
-
187
- **APA:**
188
-
189
- [More Information Needed]
190
-
191
- ## Glossary [optional]
192
-
193
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
194
-
195
- [More Information Needed]
196
-
197
- ## More Information [optional]
198
-
199
- [More Information Needed]
200
-
201
- ## Model Card Authors [optional]
202
-
203
- [More Information Needed]
204
-
205
- ## Model Card Contact
206
-
207
- [More Information Needed]
208
- ### Framework versions
209
-
210
- - PEFT 0.19.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
lora/adapter_config.json DELETED
@@ -1,52 +0,0 @@
1
- {
2
- "alora_invocation_tokens": null,
3
- "alpha_pattern": {},
4
- "arrow_config": null,
5
- "auto_mapping": {
6
- "base_model_class": "Qwen3_5ForConditionalGeneration",
7
- "parent_library": "transformers.models.qwen3_5.modeling_qwen3_5",
8
- "unsloth_fixed": true
9
- },
10
- "base_model_name_or_path": "Jackrong/Qwen3.5-9B-DeepSeek-V4-Flash",
11
- "bias": "none",
12
- "corda_config": null,
13
- "ensure_weight_tying": false,
14
- "eva_config": null,
15
- "exclude_modules": null,
16
- "fan_in_fan_out": false,
17
- "inference_mode": true,
18
- "init_lora_weights": true,
19
- "layer_replication": null,
20
- "layers_pattern": null,
21
- "layers_to_transform": null,
22
- "loftq_config": {},
23
- "lora_alpha": 64,
24
- "lora_bias": false,
25
- "lora_dropout": 0.0,
26
- "lora_ga_config": null,
27
- "megatron_config": null,
28
- "megatron_core": "megatron.core",
29
- "modules_to_save": null,
30
- "peft_type": "LORA",
31
- "peft_version": "0.19.1",
32
- "qalora_group_size": 16,
33
- "r": 32,
34
- "rank_pattern": {},
35
- "revision": null,
36
- "target_modules": [
37
- "v_proj",
38
- "k_proj",
39
- "q_proj",
40
- "down_proj",
41
- "gate_proj",
42
- "up_proj",
43
- "o_proj"
44
- ],
45
- "target_parameters": null,
46
- "task_type": "CAUSAL_LM",
47
- "trainable_token_indices": null,
48
- "use_bdlora": null,
49
- "use_dora": false,
50
- "use_qalora": false,
51
- "use_rslora": false
52
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
lora/adapter_model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8def793f603c164ee8a5c8e8f462864d34a8bc56a898aa285d390832787e5cf
3
- size 232821904
 
 
 
 
lora/chat_template.jinja DELETED
@@ -1,88 +0,0 @@
1
- {%- if tools %}
2
- {{- '<|im_start|>system\n' }}
3
- {%- if messages[0].role == 'system' %}
4
- {{- messages[0].content + '\n\n' }}
5
- {%- endif %}
6
- {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
- {%- for tool in tools %}
8
- {{- "\n" }}
9
- {{- tool | tojson }}
10
- {%- endfor %}
11
- {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
- {%- else %}
13
- {%- if messages[0].role == 'system' %}
14
- {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
- {%- endif %}
16
- {%- endif %}
17
- {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
- {%- for message in messages[::-1] %}
19
- {%- set index = (messages|length - 1) - loop.index0 %}
20
- {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
- {%- set ns.multi_step_tool = false %}
22
- {%- set ns.last_query_index = index %}
23
- {%- endif %}
24
- {%- endfor %}
25
- {%- for message in messages %}
26
- {%- if message.content is string %}
27
- {%- set content = message.content %}
28
- {%- else %}
29
- {%- set content = '' %}
30
- {%- endif %}
31
- {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
- {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
- {%- elif message.role == "assistant" %}
34
- {%- set reasoning_content = '' %}
35
- {%- if message.reasoning_content is string %}
36
- {%- set reasoning_content = message.reasoning_content %}
37
- {%- else %}
38
- {%- if '</think>' in content %}
39
- {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
- {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
- {%- endif %}
42
- {%- endif %}
43
- {%- if loop.index0 > ns.last_query_index %}
44
- {%- if loop.last or (not loop.last and reasoning_content) %}
45
- {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
- {%- else %}
47
- {{- '<|im_start|>' + message.role + '\n' + content }}
48
- {%- endif %}
49
- {%- else %}
50
- {{- '<|im_start|>' + message.role + '\n' + content }}
51
- {%- endif %}
52
- {%- if message.tool_calls %}
53
- {%- for tool_call in message.tool_calls %}
54
- {%- if (loop.first and content) or (not loop.first) %}
55
- {{- '\n' }}
56
- {%- endif %}
57
- {%- if tool_call.function %}
58
- {%- set tool_call = tool_call.function %}
59
- {%- endif %}
60
- {{- '<tool_call>\n{"name": "' }}
61
- {{- tool_call.name }}
62
- {{- '", "arguments": ' }}
63
- {%- if tool_call.arguments is string %}
64
- {{- tool_call.arguments }}
65
- {%- else %}
66
- {{- tool_call.arguments | tojson }}
67
- {%- endif %}
68
- {{- '}\n</tool_call>' }}
69
- {%- endfor %}
70
- {%- endif %}
71
- {{- '<|im_end|>\n' }}
72
- {%- elif message.role == "tool" %}
73
- {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
- {{- '<|im_start|>user' }}
75
- {%- endif %}
76
- {{- '\n<tool_response>\n' }}
77
- {{- content }}
78
- {{- '\n</tool_response>' }}
79
- {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
- {{- '<|im_end|>\n' }}
81
- {%- endif %}
82
- {%- endif %}
83
- {%- endfor %}
84
- {%- if add_generation_prompt %}
85
- {{- '<|im_start|>assistant
86
- <think>
87
- ' }}
88
- {%- endif %}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
lora/processor_config.json DELETED
@@ -1,63 +0,0 @@
1
- {
2
- "image_processor": {
3
- "data_format": "channels_first",
4
- "do_convert_rgb": true,
5
- "do_normalize": true,
6
- "do_rescale": true,
7
- "do_resize": true,
8
- "image_mean": [
9
- 0.5,
10
- 0.5,
11
- 0.5
12
- ],
13
- "image_processor_type": "Qwen2VLImageProcessor",
14
- "image_std": [
15
- 0.5,
16
- 0.5,
17
- 0.5
18
- ],
19
- "merge_size": 2,
20
- "patch_size": 16,
21
- "resample": 3,
22
- "rescale_factor": 0.00392156862745098,
23
- "size": {
24
- "longest_edge": 16777216,
25
- "shortest_edge": 65536
26
- },
27
- "temporal_patch_size": 2
28
- },
29
- "processor_class": "Qwen3VLProcessor",
30
- "video_processor": {
31
- "data_format": "channels_first",
32
- "default_to_square": true,
33
- "do_convert_rgb": true,
34
- "do_normalize": true,
35
- "do_rescale": true,
36
- "do_resize": true,
37
- "do_sample_frames": true,
38
- "fps": 2,
39
- "image_mean": [
40
- 0.5,
41
- 0.5,
42
- 0.5
43
- ],
44
- "image_std": [
45
- 0.5,
46
- 0.5,
47
- 0.5
48
- ],
49
- "max_frames": 768,
50
- "merge_size": 2,
51
- "min_frames": 4,
52
- "patch_size": 16,
53
- "resample": 3,
54
- "rescale_factor": 0.00392156862745098,
55
- "return_metadata": false,
56
- "size": {
57
- "longest_edge": 25165824,
58
- "shortest_edge": 4096
59
- },
60
- "temporal_patch_size": 2,
61
- "video_processor_type": "Qwen3VLVideoProcessor"
62
- }
63
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
lora/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:87a7830d63fcf43bf241c3c5242e96e62dd3fdc29224ca26fed8ea333db72de4
3
- size 19989343
 
 
 
 
lora/tokenizer_config.json DELETED
@@ -1,299 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "audio_bos_token": "<|audio_start|>",
4
- "audio_eos_token": "<|audio_end|>",
5
- "audio_token": "<|audio_pad|>",
6
- "backend": "tokenizers",
7
- "bos_token": null,
8
- "clean_up_tokenization_spaces": false,
9
- "eos_token": "<|im_end|>",
10
- "errors": "replace",
11
- "image_token": "<|image_pad|>",
12
- "is_local": false,
13
- "model_max_length": 262144,
14
- "model_specific_special_tokens": {
15
- "audio_bos_token": "<|audio_start|>",
16
- "audio_eos_token": "<|audio_end|>",
17
- "audio_token": "<|audio_pad|>",
18
- "image_token": "<|image_pad|>",
19
- "video_token": "<|video_pad|>",
20
- "vision_bos_token": "<|vision_start|>",
21
- "vision_eos_token": "<|vision_end|>"
22
- },
23
- "pad_token": "<|vision_pad|>",
24
- "padding_side": "right",
25
- "pretokenize_regex": "(?i:'s|'t|'re|'ve|'m|'ll|'d)|[^\\r\\n\\p{L}\\p{N}]?[\\p{L}\\p{M}]+|\\p{N}| ?[^\\s\\p{L}\\p{M}\\p{N}]+[\\r\\n]*|\\s*[\\r\\n]+|\\s+(?!\\S)|\\s+",
26
- "processor_class": "Qwen3VLProcessor",
27
- "split_special_tokens": false,
28
- "tokenizer_class": "TokenizersBackend",
29
- "unk_token": null,
30
- "video_token": "<|video_pad|>",
31
- "vision_bos_token": "<|vision_start|>",
32
- "vision_eos_token": "<|vision_end|>",
33
- "added_tokens_decoder": {
34
- "248044": {
35
- "content": "<|endoftext|>",
36
- "single_word": false,
37
- "lstrip": false,
38
- "rstrip": false,
39
- "normalized": false,
40
- "special": true
41
- },
42
- "248045": {
43
- "content": "<|im_start|>",
44
- "single_word": false,
45
- "lstrip": false,
46
- "rstrip": false,
47
- "normalized": false,
48
- "special": true
49
- },
50
- "248046": {
51
- "content": "<|im_end|>",
52
- "single_word": false,
53
- "lstrip": false,
54
- "rstrip": false,
55
- "normalized": false,
56
- "special": true
57
- },
58
- "248047": {
59
- "content": "<|object_ref_start|>",
60
- "single_word": false,
61
- "lstrip": false,
62
- "rstrip": false,
63
- "normalized": false,
64
- "special": true
65
- },
66
- "248048": {
67
- "content": "<|object_ref_end|>",
68
- "single_word": false,
69
- "lstrip": false,
70
- "rstrip": false,
71
- "normalized": false,
72
- "special": true
73
- },
74
- "248049": {
75
- "content": "<|box_start|>",
76
- "single_word": false,
77
- "lstrip": false,
78
- "rstrip": false,
79
- "normalized": false,
80
- "special": true
81
- },
82
- "248050": {
83
- "content": "<|box_end|>",
84
- "single_word": false,
85
- "lstrip": false,
86
- "rstrip": false,
87
- "normalized": false,
88
- "special": true
89
- },
90
- "248051": {
91
- "content": "<|quad_start|>",
92
- "single_word": false,
93
- "lstrip": false,
94
- "rstrip": false,
95
- "normalized": false,
96
- "special": true
97
- },
98
- "248052": {
99
- "content": "<|quad_end|>",
100
- "single_word": false,
101
- "lstrip": false,
102
- "rstrip": false,
103
- "normalized": false,
104
- "special": true
105
- },
106
- "248053": {
107
- "content": "<|vision_start|>",
108
- "single_word": false,
109
- "lstrip": false,
110
- "rstrip": false,
111
- "normalized": false,
112
- "special": true
113
- },
114
- "248054": {
115
- "content": "<|vision_end|>",
116
- "single_word": false,
117
- "lstrip": false,
118
- "rstrip": false,
119
- "normalized": false,
120
- "special": true
121
- },
122
- "248055": {
123
- "content": "<|vision_pad|>",
124
- "single_word": false,
125
- "lstrip": false,
126
- "rstrip": false,
127
- "normalized": false,
128
- "special": true
129
- },
130
- "248056": {
131
- "content": "<|image_pad|>",
132
- "single_word": false,
133
- "lstrip": false,
134
- "rstrip": false,
135
- "normalized": false,
136
- "special": true
137
- },
138
- "248057": {
139
- "content": "<|video_pad|>",
140
- "single_word": false,
141
- "lstrip": false,
142
- "rstrip": false,
143
- "normalized": false,
144
- "special": true
145
- },
146
- "248058": {
147
- "content": "<tool_call>",
148
- "single_word": false,
149
- "lstrip": false,
150
- "rstrip": false,
151
- "normalized": false,
152
- "special": false
153
- },
154
- "248059": {
155
- "content": "</tool_call>",
156
- "single_word": false,
157
- "lstrip": false,
158
- "rstrip": false,
159
- "normalized": false,
160
- "special": false
161
- },
162
- "248060": {
163
- "content": "<|fim_prefix|>",
164
- "single_word": false,
165
- "lstrip": false,
166
- "rstrip": false,
167
- "normalized": false,
168
- "special": false
169
- },
170
- "248061": {
171
- "content": "<|fim_middle|>",
172
- "single_word": false,
173
- "lstrip": false,
174
- "rstrip": false,
175
- "normalized": false,
176
- "special": false
177
- },
178
- "248062": {
179
- "content": "<|fim_suffix|>",
180
- "single_word": false,
181
- "lstrip": false,
182
- "rstrip": false,
183
- "normalized": false,
184
- "special": false
185
- },
186
- "248063": {
187
- "content": "<|fim_pad|>",
188
- "single_word": false,
189
- "lstrip": false,
190
- "rstrip": false,
191
- "normalized": false,
192
- "special": false
193
- },
194
- "248064": {
195
- "content": "<|repo_name|>",
196
- "single_word": false,
197
- "lstrip": false,
198
- "rstrip": false,
199
- "normalized": false,
200
- "special": false
201
- },
202
- "248065": {
203
- "content": "<|file_sep|>",
204
- "single_word": false,
205
- "lstrip": false,
206
- "rstrip": false,
207
- "normalized": false,
208
- "special": false
209
- },
210
- "248066": {
211
- "content": "<tool_response>",
212
- "single_word": false,
213
- "lstrip": false,
214
- "rstrip": false,
215
- "normalized": false,
216
- "special": false
217
- },
218
- "248067": {
219
- "content": "</tool_response>",
220
- "single_word": false,
221
- "lstrip": false,
222
- "rstrip": false,
223
- "normalized": false,
224
- "special": false
225
- },
226
- "248068": {
227
- "content": "<think>",
228
- "single_word": false,
229
- "lstrip": false,
230
- "rstrip": false,
231
- "normalized": false,
232
- "special": false
233
- },
234
- "248069": {
235
- "content": "</think>",
236
- "single_word": false,
237
- "lstrip": false,
238
- "rstrip": false,
239
- "normalized": false,
240
- "special": false
241
- },
242
- "248070": {
243
- "content": "<|audio_start|>",
244
- "single_word": false,
245
- "lstrip": false,
246
- "rstrip": false,
247
- "normalized": false,
248
- "special": true
249
- },
250
- "248071": {
251
- "content": "<|audio_end|>",
252
- "single_word": false,
253
- "lstrip": false,
254
- "rstrip": false,
255
- "normalized": false,
256
- "special": true
257
- },
258
- "248072": {
259
- "content": "<tts_pad>",
260
- "single_word": false,
261
- "lstrip": false,
262
- "rstrip": false,
263
- "normalized": false,
264
- "special": true
265
- },
266
- "248073": {
267
- "content": "<tts_text_bos>",
268
- "single_word": false,
269
- "lstrip": false,
270
- "rstrip": false,
271
- "normalized": false,
272
- "special": true
273
- },
274
- "248074": {
275
- "content": "<tts_text_eod>",
276
- "single_word": false,
277
- "lstrip": false,
278
- "rstrip": false,
279
- "normalized": false,
280
- "special": true
281
- },
282
- "248075": {
283
- "content": "<tts_text_bos_single>",
284
- "single_word": false,
285
- "lstrip": false,
286
- "rstrip": false,
287
- "normalized": false,
288
- "special": true
289
- },
290
- "248076": {
291
- "content": "<|audio_pad|>",
292
- "single_word": false,
293
- "lstrip": false,
294
- "rstrip": false,
295
- "normalized": false,
296
- "special": true
297
- }
298
- }
299
- }