output = llm(
"Once upon a time,",
max_tokens=512,
echo=True
)
print(output)TinySolar-248m-4k โ Q4_K_M GGUF
Quantized by the Temple AI Nervous System (TAINS) forge.
- Base model: upstage/TinySolar-248m-4k
- Parameters: 248M
- Quantization: Q4_K_M
- Neural tier: cerebellum
- File size: 148.5MB
Converted using llama.cpp convert_hf_to_gguf.py + llama-quantize.
- Downloads last month
- 22
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for kenww53/tinysolar-248m-GGUF
Base model
upstage/TinySolar-248m-4k
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="kenww53/tinysolar-248m-GGUF", filename="tinysolar-248m-Q4_K_M.gguf", )