LLaMA3

custom Llama-based model trained on tiny shakespeare text using JAX.

Model Details

  • Architecture: Llama-based Transformer
  • Parameters: ~31.6M
  • Training Framework: JAX
  • Tokenization: GPT-2 Encoding (tiktoken)

How to use

from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("parkneurals/LLaMA3")
Downloads last month
137
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support