WithIn US AI (((GGUF MODELS))
Collection
LLM MODELS TRAINED, FINE-TUNED, MERGED BY (WITHIN US AI) โข 14 items โข Updated โข 3
Llama-3.2-OctoThinker-iNano-1B-GGUF is the GGUF quantized release of the main model:
Main model repo:
https://huggingface.co/gss1147/Llama-3.2-OctoThinker-iNano-1B
This repository packages the model for efficient local inference in GGUF-compatible runtimes such as llama.cpp, LM Studio, and similar local tools.
This GGUF repository corresponds to the main model repo:
gss1147/Llama-3.2-OctoThinker-iNano-1B
If you want the original non-GGUF model, training/merge details, tokenizer files, and main repository metadata, use the repo above.
This GGUF repository currently includes:
This model is intended for:
Choose the file that best matches your hardware:
llama-cli -m /path/to/Llama-3.2-OctoThinker-iNano-1B.Q4_K_M.gguf -p "Explain recursion in Python with a simple example."
4-bit
5-bit
16-bit
Base model
WithinUsAI/Llama3.2-OctoThinker-iNano-1B