elliotthwang commited on
Commit
a98820a
·
verified ·
1 Parent(s): 94b8e06

elliotthwang/Mistral-7B-Instruct-v0.2-tw-train_ouputs

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: elliotthwang/Ministral-4b-instruct-tw_1k
3
  library_name: transformers
4
  model_name: outputs
5
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for outputs
13
 
14
- This model is a fine-tuned version of [elliotthwang/Ministral-4b-instruct-tw_1k](https://huggingface.co/elliotthwang/Ministral-4b-instruct-tw_1k).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -34,10 +34,10 @@ This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.16.1
38
- - Transformers: 4.51.1
39
  - Pytorch: 2.6.0+cu124
40
- - Datasets: 3.5.0
41
  - Tokenizers: 0.21.1
42
 
43
  ## Citations
@@ -49,7 +49,7 @@ Cite TRL as:
49
  ```bibtex
50
  @misc{vonwerra2022trl,
51
  title = {{TRL: Transformer Reinforcement Learning}},
52
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
53
  year = 2020,
54
  journal = {GitHub repository},
55
  publisher = {GitHub},
 
1
  ---
2
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
3
  library_name: transformers
4
  model_name: outputs
5
  tags:
 
11
 
12
  # Model Card for outputs
13
 
14
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.18.1
38
+ - Transformers: 4.52.2
39
  - Pytorch: 2.6.0+cu124
40
+ - Datasets: 3.6.0
41
  - Tokenizers: 0.21.1
42
 
43
  ## Citations
 
49
  ```bibtex
50
  @misc{vonwerra2022trl,
51
  title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
  year = 2020,
54
  journal = {GitHub repository},
55
  publisher = {GitHub},
adapter_config.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "elliotthwang/Ministral-4b-instruct-tw_1k",
5
  "bias": "none",
 
6
  "eva_config": null,
7
  "exclude_modules": null,
8
  "fan_in_fan_out": false,
@@ -24,12 +25,13 @@
24
  "revision": null,
25
  "target_modules": [
26
  "o_proj",
27
- "gate_proj",
28
- "q_proj",
29
  "v_proj",
30
- "k_proj"
 
31
  ],
32
  "task_type": "CAUSAL_LM",
 
33
  "use_dora": false,
34
  "use_rslora": false
35
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
5
  "bias": "none",
6
+ "corda_config": null,
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
 
25
  "revision": null,
26
  "target_modules": [
27
  "o_proj",
28
+ "k_proj",
 
29
  "v_proj",
30
+ "gate_proj",
31
+ "q_proj"
32
  ],
33
  "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
  "use_dora": false,
36
  "use_rslora": false
37
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aacb936da69c016bd8269af3f4779c532230c31c50434b0eecf02d3db87e6d9a
3
- size 38839760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52189c532977c742b7fe78a336a35b654c9e4470ce77f75b0a20f31ef6cc135f
3
+ size 46179856
chat_template.jinja ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if messages[0]['role'] == 'system' %}
2
+ {%- set system_message = messages[0]['content'] %}
3
+ {%- set loop_messages = messages[1:] %}
4
+ {%- else %}
5
+ {%- set loop_messages = messages %}
6
+ {%- endif %}
7
+
8
+ {{- bos_token }}
9
+ {%- for message in loop_messages %}
10
+ {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}
11
+ {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}
12
+ {%- endif %}
13
+ {%- if message['role'] == 'user' %}
14
+ {%- if loop.first and system_message is defined %}
15
+ {{- ' [INST] ' + system_message + '\n\n' + message['content'] + ' [/INST]' }}
16
+ {%- else %}
17
+ {{- ' [INST] ' + message['content'] + ' [/INST]' }}
18
+ {%- endif %}
19
+ {%- elif message['role'] == 'assistant' %}
20
+ {{- ' ' + message['content'] + eos_token}}
21
+ {%- else %}
22
+ {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}
23
+ {%- endif %}
24
+ {%- endfor %}
runs/May30_10-14-38_2a9f8f30fcd1/events.out.tfevents.1748600080.2a9f8f30fcd1.1234.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f12c84ff1231c1288a827061b5c0e71a3ef9320ea0f1d0ff5e5b86bdacf0c704
3
+ size 335322
special_tokens_map.json CHANGED
@@ -13,13 +13,6 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
 
 
 
 
 
 
16
  "unk_token": {
17
  "content": "<unk>",
18
  "lstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
- "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
@@ -30,14 +30,12 @@
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
33
- "chat_template": "{{ bos_token }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '### User:\\n' + message['content'] + '\\n#### Assistant:\\n' }}{% elif message['role'] == 'system' %}{{ '### System:\\n' + message['content'] + '\\n' }}{% elif message['role'] == 'assistant' %}{{ message['content'] }}{% endif %}{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
  "extra_special_tokens": {},
37
  "legacy": false,
38
  "model_max_length": 1000000000000000019884624838656,
39
- "pad_token": "</s>",
40
- "padding_side": "right",
41
  "sp_model_kwargs": {},
42
  "spaces_between_special_tokens": false,
43
  "tokenizer_class": "LlamaTokenizer",
 
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
 
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
 
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "extra_special_tokens": {},
36
  "legacy": false,
37
  "model_max_length": 1000000000000000019884624838656,
38
+ "pad_token": null,
 
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
41
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82404c1bf537fc27d98e14da82fbebf16f5c7bb3d532cadb36de4fc1fba53217
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:730426600dec18a000cbbe9011a0b4e35d040a9bdc822fcbf7ca8161a998bd37
3
  size 5560