RecursiveMAS commited on
Commit
364f2ce
·
verified ·
1 Parent(s): cc76af9

Clear old files before upload

Browse files
README.md DELETED
@@ -1,7 +0,0 @@
1
- ---
2
- base_model:
3
- - meta-llama/Llama-3.2-1B-Instruct
4
- license: mit
5
- ---
6
-
7
- The Critic Agent of Sequential-Style (light) MAS in the paper: Recursive Multi-Agent Systems. (base model: meta-llama/Llama-3.2-1B-Instruct)
 
 
 
 
 
 
 
 
adapter(code).pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5053c60fed06d388b423bbb2ef0677af8f160a146789e790e1c5ed1e28f18844
3
- size 16805005
 
 
 
 
adapter(math).pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4988907ff79262e4843940f358544a6079c54d0016f680dbb9dc9cf15a870f49
3
- size 16805005
 
 
 
 
adapter_config.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "adapter_type": "ln_res_adapter"
3
- }
 
 
 
 
chat_template.jinja DELETED
@@ -1,93 +0,0 @@
1
- {{- bos_token }}
2
- {%- if custom_tools is defined %}
3
- {%- set tools = custom_tools %}
4
- {%- endif %}
5
- {%- if not tools_in_user_message is defined %}
6
- {%- set tools_in_user_message = true %}
7
- {%- endif %}
8
- {%- if not date_string is defined %}
9
- {%- if strftime_now is defined %}
10
- {%- set date_string = strftime_now("%d %b %Y") %}
11
- {%- else %}
12
- {%- set date_string = "26 Jul 2024" %}
13
- {%- endif %}
14
- {%- endif %}
15
- {%- if not tools is defined %}
16
- {%- set tools = none %}
17
- {%- endif %}
18
-
19
- {#- This block extracts the system message, so we can slot it into the right place. #}
20
- {%- if messages[0]['role'] == 'system' %}
21
- {%- set system_message = messages[0]['content']|trim %}
22
- {%- set messages = messages[1:] %}
23
- {%- else %}
24
- {%- set system_message = "" %}
25
- {%- endif %}
26
-
27
- {#- System message #}
28
- {{- "<|start_header_id|>system<|end_header_id|>\n\n" }}
29
- {%- if tools is not none %}
30
- {{- "Environment: ipython\n" }}
31
- {%- endif %}
32
- {{- "Cutting Knowledge Date: December 2023\n" }}
33
- {{- "Today Date: " + date_string + "\n\n" }}
34
- {%- if tools is not none and not tools_in_user_message %}
35
- {{- "You have access to the following functions. To call a function, please respond with JSON for a function call." }}
36
- {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }}
37
- {{- "Do not use variables.\n\n" }}
38
- {%- for t in tools %}
39
- {{- t | tojson(indent=4) }}
40
- {{- "\n\n" }}
41
- {%- endfor %}
42
- {%- endif %}
43
- {{- system_message }}
44
- {{- "<|eot_id|>" }}
45
-
46
- {#- Custom tools are passed in a user message with some extra guidance #}
47
- {%- if tools_in_user_message and not tools is none %}
48
- {#- Extract the first user message so we can plug it in here #}
49
- {%- if messages | length != 0 %}
50
- {%- set first_user_message = messages[0]['content']|trim %}
51
- {%- set messages = messages[1:] %}
52
- {%- else %}
53
- {{- raise_exception("Cannot put tools in the first user message when there's no first user message!") }}
54
- {%- endif %}
55
- {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}
56
- {{- "Given the following functions, please respond with a JSON for a function call " }}
57
- {{- "with its proper arguments that best answers the given prompt.\n\n" }}
58
- {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }}
59
- {{- "Do not use variables.\n\n" }}
60
- {%- for t in tools %}
61
- {{- t | tojson(indent=4) }}
62
- {{- "\n\n" }}
63
- {%- endfor %}
64
- {{- first_user_message + "<|eot_id|>"}}
65
- {%- endif %}
66
-
67
- {%- for message in messages %}
68
- {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}
69
- {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}
70
- {%- elif 'tool_calls' in message %}
71
- {%- if not message.tool_calls|length == 1 %}
72
- {{- raise_exception("This model only supports single tool-calls at once!") }}
73
- {%- endif %}
74
- {%- set tool_call = message.tool_calls[0].function %}
75
- {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}
76
- {{- '{"name": "' + tool_call.name + '", ' }}
77
- {{- '"parameters": ' }}
78
- {{- tool_call.arguments | tojson }}
79
- {{- "}" }}
80
- {{- "<|eot_id|>" }}
81
- {%- elif message.role == "tool" or message.role == "ipython" %}
82
- {{- "<|start_header_id|>ipython<|end_header_id|>\n\n" }}
83
- {%- if message.content is mapping or message.content is iterable %}
84
- {{- message.content | tojson }}
85
- {%- else %}
86
- {{- message.content }}
87
- {%- endif %}
88
- {{- "<|eot_id|>" }}
89
- {%- endif %}
90
- {%- endfor %}
91
- {%- if add_generation_prompt %}
92
- {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}
93
- {%- endif %}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "architectures": [
3
- "LlamaForCausalLM"
4
- ],
5
- "attention_bias": false,
6
- "attention_dropout": 0.0,
7
- "bos_token_id": 128000,
8
- "dtype": "bfloat16",
9
- "eos_token_id": [
10
- 128001,
11
- 128008,
12
- 128009
13
- ],
14
- "head_dim": 64,
15
- "hidden_act": "silu",
16
- "hidden_size": 2048,
17
- "initializer_range": 0.02,
18
- "intermediate_size": 8192,
19
- "max_position_embeddings": 131072,
20
- "mlp_bias": false,
21
- "model_type": "llama",
22
- "num_attention_heads": 32,
23
- "num_hidden_layers": 16,
24
- "num_key_value_heads": 8,
25
- "pad_token_id": null,
26
- "pretraining_tp": 1,
27
- "rms_norm_eps": 1e-05,
28
- "rope_parameters": {
29
- "factor": 32.0,
30
- "high_freq_factor": 4.0,
31
- "low_freq_factor": 1.0,
32
- "original_max_position_embeddings": 8192,
33
- "rope_theta": 500000.0,
34
- "rope_type": "llama3"
35
- },
36
- "tie_word_embeddings": true,
37
- "transformers_version": "5.3.0",
38
- "use_cache": true,
39
- "vocab_size": 128256
40
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_config.json DELETED
@@ -1,12 +0,0 @@
1
- {
2
- "bos_token_id": 128000,
3
- "do_sample": true,
4
- "eos_token_id": [
5
- 128001,
6
- 128008,
7
- 128009
8
- ],
9
- "temperature": 0.6,
10
- "top_p": 0.9,
11
- "transformers_version": "5.3.0"
12
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
innerlink_config.json DELETED
@@ -1,36 +0,0 @@
1
- {
2
- "format_version": 1,
3
- "paradigm": "sequential",
4
- "variant": "light",
5
- "role": "Critic",
6
- "backbone": "Llama3.2-1B",
7
- "shared_files": [
8
- "adapter_config.json",
9
- "chat_template.jinja",
10
- "config.json",
11
- "generation_config.json",
12
- "model.safetensors",
13
- "tokenizer.json",
14
- "tokenizer_config.json"
15
- ],
16
- "tasks": {
17
- "math": {
18
- "adapter.pt": "adapter(math).pt",
19
- "adapter_config.json": "adapter_config.json",
20
- "chat_template.jinja": "chat_template.jinja",
21
- "config.json": "config.json",
22
- "generation_config.json": "generation_config.json",
23
- "tokenizer.json": "tokenizer.json",
24
- "tokenizer_config.json": "tokenizer_config.json"
25
- },
26
- "code": {
27
- "adapter.pt": "adapter(code).pt",
28
- "adapter_config.json": "adapter_config.json",
29
- "chat_template.jinja": "chat_template.jinja",
30
- "config.json": "config.json",
31
- "generation_config.json": "generation_config.json",
32
- "tokenizer.json": "tokenizer.json",
33
- "tokenizer_config.json": "tokenizer_config.json"
34
- }
35
- }
36
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ff795ff6a07e6a68085d206fb84417da2f083f68391c2843cd2b8ac6df8538f
3
- size 2471645608
 
 
 
 
tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
- size 17209920
 
 
 
 
tokenizer_config.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "backend": "tokenizers",
3
- "bos_token": "<|begin_of_text|>",
4
- "clean_up_tokenization_spaces": true,
5
- "eos_token": "<|eot_id|>",
6
- "is_local": false,
7
- "model_input_names": [
8
- "input_ids",
9
- "attention_mask"
10
- ],
11
- "model_max_length": 131072,
12
- "pad_token": "<|eot_id|>",
13
- "tokenizer_class": "TokenizersBackend"
14
- }