Upload folder using huggingface_hub
Browse files- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/aqua.npy +3 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/boolq.npy +3 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/collect.txt +30 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/commonsenseqa.npy +3 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/gsm8k.npy +3 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/meta.json +51 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/openbookqa.npy +3 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/piqa.npy +3 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/qasc.npy +3 -0
- artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/strategyqa.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/aqua.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/boolq.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/collect.txt +30 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/commonsenseqa.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/gsm8k.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/meta.json +51 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/openbookqa.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/piqa.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/qasc.npy +3 -0
- artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/strategyqa.npy +3 -0
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/aqua.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c1c570097c0685a8fe3210e081d1e63c2dc621800427903ad9b42be043da76b6
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/boolq.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:332b0c115c6a1997fa33dc126d367d2007585704eb94a25e0251d0180abf73f9
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/collect.txt
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[Cmd] collect_decode_acts.py --model Qwen/Qwen2.5-7B-Instruct --device cuda --model_dtype fp32 --layer 10 --n_prompts 128 --calib_max_new_tokens 256 --max_prompt_len 512 --per_task_max_states 20000 --batch_size 4 --seed 42 --out_dir results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42 --save_dtype fp32 --out_txt results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/collect.txt --overwrite
|
| 2 |
+
[Data] tasks=['gsm8k', 'commonsenseqa', 'strategyqa', 'aqua', 'openbookqa', 'qasc', 'boolq', 'piqa']
|
| 3 |
+
[Data] task=gsm8k prompts=128
|
| 4 |
+
[Data] task=commonsenseqa prompts=128
|
| 5 |
+
[Data] task=strategyqa prompts=128
|
| 6 |
+
[Data] task=aqua prompts=128
|
| 7 |
+
[Data] task=openbookqa prompts=128
|
| 8 |
+
[Data] task=qasc prompts=128
|
| 9 |
+
[Data] task=boolq prompts=128
|
| 10 |
+
[Data] task=piqa prompts=128
|
| 11 |
+
[Collect] task=gsm8k
|
| 12 |
+
[Collect] task=commonsenseqa
|
| 13 |
+
[Collect] task=strategyqa
|
| 14 |
+
[Collect] task=aqua
|
| 15 |
+
[Collect] task=openbookqa
|
| 16 |
+
[Collect] task=qasc
|
| 17 |
+
[Collect] task=boolq
|
| 18 |
+
[Collect] task=piqa
|
| 19 |
+
[Collect] task=gsm8k raw_states=32768 x 3584
|
| 20 |
+
[Collect] task=commonsenseqa raw_states=32768 x 3584
|
| 21 |
+
[Collect] task=strategyqa raw_states=32768 x 3584
|
| 22 |
+
[Collect] task=aqua raw_states=32768 x 3584
|
| 23 |
+
[Collect] task=openbookqa raw_states=32768 x 3584
|
| 24 |
+
[Collect] task=qasc raw_states=32768 x 3584
|
| 25 |
+
[Collect] task=boolq raw_states=32768 x 3584
|
| 26 |
+
[Collect] task=piqa raw_states=32768 x 3584
|
| 27 |
+
[Fair] balanced states per task = 20000
|
| 28 |
+
[Done]
|
| 29 |
+
Saved meta: results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/meta.json
|
| 30 |
+
Saved npy : results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/*.npy
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/commonsenseqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6e19450db7ecaa63224cf42126e05d80e19a2d35a40d2c77a27cd015d539f716
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/gsm8k.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df7e4ccf8afafcf8efb989d498ca25abe0549ae917c6e61aa63b6cb593a5e231
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/meta.json
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model": "Qwen/Qwen2.5-7B-Instruct",
|
| 4 |
+
"device": "cuda",
|
| 5 |
+
"model_dtype": "fp32",
|
| 6 |
+
"layer": 10,
|
| 7 |
+
"n_prompts": 128,
|
| 8 |
+
"max_prompt_len": 512,
|
| 9 |
+
"calib_max_new_tokens": 256,
|
| 10 |
+
"calib_decoding": "greedy",
|
| 11 |
+
"temperature": 0.7,
|
| 12 |
+
"top_p": 0.9,
|
| 13 |
+
"top_k": 0,
|
| 14 |
+
"batch_size": 4,
|
| 15 |
+
"per_task_max_states": 20000,
|
| 16 |
+
"balance_to": "min",
|
| 17 |
+
"seed": 42,
|
| 18 |
+
"save_dtype": "fp32"
|
| 19 |
+
},
|
| 20 |
+
"tasks": [
|
| 21 |
+
"gsm8k",
|
| 22 |
+
"commonsenseqa",
|
| 23 |
+
"strategyqa",
|
| 24 |
+
"aqua",
|
| 25 |
+
"openbookqa",
|
| 26 |
+
"qasc",
|
| 27 |
+
"boolq",
|
| 28 |
+
"piqa"
|
| 29 |
+
],
|
| 30 |
+
"raw_counts": {
|
| 31 |
+
"gsm8k": 32768,
|
| 32 |
+
"commonsenseqa": 32768,
|
| 33 |
+
"strategyqa": 32768,
|
| 34 |
+
"aqua": 32768,
|
| 35 |
+
"openbookqa": 32768,
|
| 36 |
+
"qasc": 32768,
|
| 37 |
+
"boolq": 32768,
|
| 38 |
+
"piqa": 32768
|
| 39 |
+
},
|
| 40 |
+
"balanced_states_per_task": 20000,
|
| 41 |
+
"files": {
|
| 42 |
+
"gsm8k": "gsm8k.npy",
|
| 43 |
+
"commonsenseqa": "commonsenseqa.npy",
|
| 44 |
+
"strategyqa": "strategyqa.npy",
|
| 45 |
+
"aqua": "aqua.npy",
|
| 46 |
+
"openbookqa": "openbookqa.npy",
|
| 47 |
+
"qasc": "qasc.npy",
|
| 48 |
+
"boolq": "boolq.npy",
|
| 49 |
+
"piqa": "piqa.npy"
|
| 50 |
+
}
|
| 51 |
+
}
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/openbookqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c5586d1e23630b811935d72bb2e13540d5395292663aa12d2876d5e9471be179
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/piqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ad5001af072f9ec2905e2b721423135bb2d11ba3ecbb45a0e0c50915bd47b8a
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/qasc.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ba8669e27778d422a89839b2335090d54e6e9b8d551b78ee57dfa53b4782681
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/Qwen_Qwen2.5-7B-Instruct/layer10_n128_new256_maxlen512_states20000_seed42/strategyqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad7406cf57b0409cc764e69dac4a59958496fba917fff3314fe3a3a3d8af944f
|
| 3 |
+
size 286720128
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/aqua.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f298d287b81536c351a9ed4a89dc6dd7b291514ae9e2ff4171993b55eca01467
|
| 3 |
+
size 304087168
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/boolq.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:65a25e641a96f1e1b1e5a0d4727c1996b0c3f7770d1fc13f79f5f8c0022d0da0
|
| 3 |
+
size 304087168
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/collect.txt
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[Cmd] collect_decode_acts.py --model meta-llama/Llama-2-7b-chat-hf --device cuda --model_dtype fp32 --layer 10 --n_prompts 128 --calib_max_new_tokens 256 --max_prompt_len 512 --per_task_max_states 20000 --batch_size 4 --seed 42 --out_dir results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42 --save_dtype fp32 --out_txt results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/collect.txt --overwrite
|
| 2 |
+
[Data] tasks=['gsm8k', 'commonsenseqa', 'strategyqa', 'aqua', 'openbookqa', 'qasc', 'boolq', 'piqa']
|
| 3 |
+
[Data] task=gsm8k prompts=128
|
| 4 |
+
[Data] task=commonsenseqa prompts=128
|
| 5 |
+
[Data] task=strategyqa prompts=128
|
| 6 |
+
[Data] task=aqua prompts=128
|
| 7 |
+
[Data] task=openbookqa prompts=128
|
| 8 |
+
[Data] task=qasc prompts=128
|
| 9 |
+
[Data] task=boolq prompts=128
|
| 10 |
+
[Data] task=piqa prompts=128
|
| 11 |
+
[Collect] task=gsm8k
|
| 12 |
+
[Collect] task=commonsenseqa
|
| 13 |
+
[Collect] task=strategyqa
|
| 14 |
+
[Collect] task=aqua
|
| 15 |
+
[Collect] task=openbookqa
|
| 16 |
+
[Collect] task=qasc
|
| 17 |
+
[Collect] task=boolq
|
| 18 |
+
[Collect] task=piqa
|
| 19 |
+
[Collect] task=gsm8k raw_states=18560 x 4096
|
| 20 |
+
[Collect] task=commonsenseqa raw_states=21001 x 4096
|
| 21 |
+
[Collect] task=strategyqa raw_states=23509 x 4096
|
| 22 |
+
[Collect] task=aqua raw_states=24667 x 4096
|
| 23 |
+
[Collect] task=openbookqa raw_states=20494 x 4096
|
| 24 |
+
[Collect] task=qasc raw_states=25433 x 4096
|
| 25 |
+
[Collect] task=boolq raw_states=22885 x 4096
|
| 26 |
+
[Collect] task=piqa raw_states=21000 x 4096
|
| 27 |
+
[Fair] balanced states per task = 18560
|
| 28 |
+
[Done]
|
| 29 |
+
Saved meta: results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/meta.json
|
| 30 |
+
Saved npy : results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/*.npy
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/commonsenseqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:604bf6b8d97ad18b806f92a90e3961f176ae11339b3e8586414bac04f0d1f7f9
|
| 3 |
+
size 304087168
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/gsm8k.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:289c13700e324287c1d6a7dfbb0d2a1fa04a552a142bd6e84fa21708043e1925
|
| 3 |
+
size 304087168
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/meta.json
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"config": {
|
| 3 |
+
"model": "meta-llama/Llama-2-7b-chat-hf",
|
| 4 |
+
"device": "cuda",
|
| 5 |
+
"model_dtype": "fp32",
|
| 6 |
+
"layer": 10,
|
| 7 |
+
"n_prompts": 128,
|
| 8 |
+
"max_prompt_len": 512,
|
| 9 |
+
"calib_max_new_tokens": 256,
|
| 10 |
+
"calib_decoding": "greedy",
|
| 11 |
+
"temperature": 0.7,
|
| 12 |
+
"top_p": 0.9,
|
| 13 |
+
"top_k": 0,
|
| 14 |
+
"batch_size": 4,
|
| 15 |
+
"per_task_max_states": 20000,
|
| 16 |
+
"balance_to": "min",
|
| 17 |
+
"seed": 42,
|
| 18 |
+
"save_dtype": "fp32"
|
| 19 |
+
},
|
| 20 |
+
"tasks": [
|
| 21 |
+
"gsm8k",
|
| 22 |
+
"commonsenseqa",
|
| 23 |
+
"strategyqa",
|
| 24 |
+
"aqua",
|
| 25 |
+
"openbookqa",
|
| 26 |
+
"qasc",
|
| 27 |
+
"boolq",
|
| 28 |
+
"piqa"
|
| 29 |
+
],
|
| 30 |
+
"raw_counts": {
|
| 31 |
+
"gsm8k": 18560,
|
| 32 |
+
"commonsenseqa": 21001,
|
| 33 |
+
"strategyqa": 23509,
|
| 34 |
+
"aqua": 24667,
|
| 35 |
+
"openbookqa": 20494,
|
| 36 |
+
"qasc": 25433,
|
| 37 |
+
"boolq": 22885,
|
| 38 |
+
"piqa": 21000
|
| 39 |
+
},
|
| 40 |
+
"balanced_states_per_task": 18560,
|
| 41 |
+
"files": {
|
| 42 |
+
"gsm8k": "gsm8k.npy",
|
| 43 |
+
"commonsenseqa": "commonsenseqa.npy",
|
| 44 |
+
"strategyqa": "strategyqa.npy",
|
| 45 |
+
"aqua": "aqua.npy",
|
| 46 |
+
"openbookqa": "openbookqa.npy",
|
| 47 |
+
"qasc": "qasc.npy",
|
| 48 |
+
"boolq": "boolq.npy",
|
| 49 |
+
"piqa": "piqa.npy"
|
| 50 |
+
}
|
| 51 |
+
}
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/openbookqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8a5b094d1efc9c7929d4b218ea6bfbe4df3b072d26ecefb1a1225e97681c74ef
|
| 3 |
+
size 304087168
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/piqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bb16b852f8e33147c56c183f6e3a2e23ce9aac8403fc6e8d30fbe0bb3988341e
|
| 3 |
+
size 304087168
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/qasc.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:248557735b4175fa06f50ce2370884f2cb8785a6c4c0e977db619ff29735b24e
|
| 3 |
+
size 304087168
|
artifacts/Hype1/results/acts/meta-llama_Llama-2-7b-chat-hf/layer10_n128_new256_maxlen512_states20000_seed42/strategyqa.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b5635ce8b379d9d9b88d17b10979dd9d54e25ec288d74082db1c54c0a4b05d7
|
| 3 |
+
size 304087168
|