Upload folder using huggingface_hub
Browse files- config.json +3 -3
- model-00001-of-00006.safetensors +2 -2
- model-00002-of-00006.safetensors +2 -2
- model-00003-of-00006.safetensors +1 -1
- model-00004-of-00006.safetensors +1 -1
- model-00005-of-00006.safetensors +1 -1
- model-00006-of-00006.safetensors +2 -2
- model.safetensors.index.json +2 -2
- vqvae_model_e64_k9981_L4_mse0.12622935775863497.pth +3 -0
config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "/home/quinlanp/scratch/
|
| 3 |
"architectures": [
|
| 4 |
"LlamaForCausalLM"
|
| 5 |
],
|
|
@@ -11,7 +11,7 @@
|
|
| 11 |
"hidden_size": 4096,
|
| 12 |
"initializer_range": 0.02,
|
| 13 |
"intermediate_size": 11008,
|
| 14 |
-
"max_position_embeddings":
|
| 15 |
"model_type": "llama",
|
| 16 |
"num_attention_heads": 32,
|
| 17 |
"num_hidden_layers": 32,
|
|
@@ -24,5 +24,5 @@
|
|
| 24 |
"torch_dtype": "float32",
|
| 25 |
"transformers_version": "4.36.2",
|
| 26 |
"use_cache": false,
|
| 27 |
-
"vocab_size":
|
| 28 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "/home/quinlanp/scratch/llama7b-chat-hf",
|
| 3 |
"architectures": [
|
| 4 |
"LlamaForCausalLM"
|
| 5 |
],
|
|
|
|
| 11 |
"hidden_size": 4096,
|
| 12 |
"initializer_range": 0.02,
|
| 13 |
"intermediate_size": 11008,
|
| 14 |
+
"max_position_embeddings": 4096,
|
| 15 |
"model_type": "llama",
|
| 16 |
"num_attention_heads": 32,
|
| 17 |
"num_hidden_layers": 32,
|
|
|
|
| 24 |
"torch_dtype": "float32",
|
| 25 |
"transformers_version": "4.36.2",
|
| 26 |
"use_cache": false,
|
| 27 |
+
"vocab_size": 41984
|
| 28 |
}
|
model-00001-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0947132909fbda351b179c60d30e949d973a19e604d6806ba7d3b7dbd371afe2
|
| 3 |
+
size 4936865296
|
model-00002-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f06d5bf4e0f09dce543133f560718325355bdcd240d42458855b00fdf109a6b
|
| 3 |
+
size 4924315832
|
model-00003-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4857206904
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1dd72d44842c5efaf5bc5d4cc6ff66aba051d3b4782c416140e9f79c3ee8c8c8
|
| 3 |
size 4857206904
|
model-00004-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4857206904
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cb0f8b11630c0c161abd279490547203ae704ac0e99a4a3afb2d117c8f3f61d8
|
| 3 |
size 4857206904
|
model-00005-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4857206904
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0f0f64a93007d5347c98990f8637e319a3eeb26a31add248a20c9fa2d79b3e5
|
| 3 |
size 4857206904
|
model-00006-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7c18ce28ad5d521e7e5c7f631e5f637df40d7bcbe3b20630beb5fde3c2287323
|
| 3 |
+
size 2848049976
|
model.safetensors.index.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
-
"total_size":
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"lm_head.weight": "model-00006-of-00006.safetensors",
|
|
@@ -254,7 +254,7 @@
|
|
| 254 |
"model.layers.5.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
|
| 255 |
"model.layers.5.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
|
| 256 |
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
|
| 257 |
-
"model.layers.5.self_attn.o_proj.weight": "model-
|
| 258 |
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
|
| 259 |
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
|
| 260 |
"model.layers.6.input_layernorm.weight": "model-00002-of-00006.safetensors",
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_size": 27280818176
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"lm_head.weight": "model-00006-of-00006.safetensors",
|
|
|
|
| 254 |
"model.layers.5.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
|
| 255 |
"model.layers.5.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
|
| 256 |
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
|
| 257 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
|
| 258 |
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
|
| 259 |
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
|
| 260 |
"model.layers.6.input_layernorm.weight": "model-00002-of-00006.safetensors",
|
vqvae_model_e64_k9981_L4_mse0.12622935775863497.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3356f7ef467ac15990aa21640ce9b3d402abf44bcb8d704bc379837c55f301d6
|
| 3 |
+
size 3031975
|