SmolLM-135M-Instruct-manim - Fine-tuned

This repository contains three variants of the model:

Training

  • Base model: HuggingFaceTB/SmolLM-135M-Instruct
  • Dataset: generaleoley/manim-codegen
  • Method: LoRA fine-tuning with Unsloth

Quantizations

We provide f16, bf16, f32, and q8_0 GGUF files for llama.cpp / Ollama.

Usage Example

from transformers import AutoModelForCausalLM, AutoTokenizer

tok = AutoTokenizer.from_pretrained("ArunKr/SmolLM-135M-Instruct-manim-16bit")
model = AutoModelForCausalLM.from_pretrained("ArunKr/SmolLM-135M-Instruct-manim-16bit")
print(model.generate(**tok("Hello", return_tensors="pt")))

Ollama Example

ollama run ArunKr/SmolLM-135M-Instruct-manim-gguf:<file_name>.gguf

www.ollama.com

Downloads last month
46
GGUF
Model size
0.1B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ArunKr/SmolLM-135M-Instruct-manim-gguf

Adapter
(17)
this model

Dataset used to train ArunKr/SmolLM-135M-Instruct-manim-gguf