Small Models
Collection
A list of all small models (=<1B) that I have published. • 9 items • Updated
This model is a fine-tuned version of allenai/OLMo-2-0425-1B-Instruct optimized for math tasks. It was trained using a subset of OpenAI's openai/gsm8k to improve logical reasoning and syntax accuracy.
The following hyperparameters were used during training:
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_id = "Fu01978/OLMo-2-1B-openai-gsm8k"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto")
prompt = "def find_primes(n):"
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Base model
allenai/OLMo-2-0425-1B