parkneurals commited on
Commit
3eb9f13
·
verified ·
1 Parent(s): 8569789

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +21 -0
  2. config.json +16 -0
  3. model.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # LLaMA3
3
+
4
+ custom Llama-based model trained on tiny shakespeare text using JAX.
5
+
6
+ ## Model Details
7
+ - **Architecture**: Llama-based Transformer
8
+ - **Parameters**: ~31.6M
9
+ - **Training Framework**: JAX
10
+ - **Tokenization**: GPT-2 Encoding (tiktoken)
11
+
12
+ ## How to use
13
+ ```python
14
+ from transformers import AutoModelForCausalLM, AutoTokenizer
15
+
16
+ model = AutoModelForCausalLM.from_pretrained("parkneurals/LLaMA3")
17
+ # Note: You will need a GPT-2 compatible tokenizer to run this
18
+ ```
19
+
20
+ ## Dataset
21
+ Trained on the classic Tiny Shakespeare dataset.
config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "model_type": "llama",
6
+ "hidden_size": 256,
7
+ "intermediate_size": 1024,
8
+ "num_attention_heads": 8,
9
+ "num_hidden_layers": 6,
10
+ "num_key_value_heads": 4,
11
+ "vocab_size": 50257,
12
+ "rms_norm_eps": 1e-06,
13
+ "max_position_embeddings": 512,
14
+ "torch_dtype": "float32",
15
+ "transformers_version": "4.31.0"
16
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:265298accb76bcbf89eaa3418b001117f96cf8e32462a50474f4d336a1b04fab
3
+ size 110798056