--- license: other base_model: - Qwen/Qwen3.5-4B library_name: llama.cpp tags: - gguf - qwen - qwen3.5 - code - coder - conversational - text-generation - withinusai language: - en datasets: - WithinUsAI/Python_GOD_Coder_50k - reedmayhew/gemini-3.1-pro-2048-reasoning-1100x - m-a-p/Code-Feedback - crownelius/Opus-4.6-Reasoning-2100x-formatted - crownelius/Opus4.6-No-Reasoning-260x - crownelius/Creative_Writing_Multiturn_Enhanced - HuggingFaceH4/llava-instruct-mix-vsft - Roman1111111/gemini-3-pro-10000x-hard-high-reasoning model_type: gguf inference: false --- # WithIn-Us-Coder-4B.gguf **WithIn-Us-Coder-4B.gguf** is a GGUF release from **WithIn Us AI**, built for local inference and coding-focused assistant use cases. It is based on **Qwen/Qwen3.5-4B** and distributed in quantized GGUF formats for efficient deployment in llama.cpp-compatible runtimes. ## Model Summary This model is intended as a coding-oriented conversational assistant with emphasis on: - code generation - code reasoning - implementation planning - debugging assistance - instruction following - general assistant-style chat for development workflows This repository currently provides the following GGUF variants: - `WithIn-Us-Coder-4B.Q4_K_M.gguf` - `WithIn-Us-Coder-4B.Q5_K_M.gguf` ## Creator **WithIn Us AI** is the creator of this model release, including the model packaging, fine-tuning / merging concept, process, naming, and GGUF distribution. ## Base Model This model is based on: - **Qwen/Qwen3.5-4B** Credit and appreciation go to the original creators of the base LLM architecture and weights. ## Training Data The current repository metadata lists the following datasets as part of the model’s training / fine-tuning lineage: - `WithinUsAI/Python_GOD_Coder_50k` - `reedmayhew/gemini-3.1-pro-2048-reasoning-1100x` - `m-a-p/Code-Feedback` - `crownelius/Opus-4.6-Reasoning-2100x-formatted` - `crownelius/Opus4.6-No-Reasoning-260x` - `crownelius/Creative_Writing_Multiturn_Enhanced` - `HuggingFaceH4/llava-instruct-mix-vsft` - `Roman1111111/gemini-3-pro-10000x-hard-high-reasoning` **Attribution note:** WithIn Us AI does not claim ownership over third-party base models or third-party datasets. Full credit, thanks, and attribution belong to the original model and dataset creators. ## Intended Use This model is intended for: - local coding assistants - offline development help - code explanation - bug-fixing support - prompt-based code generation - experimentation in llama.cpp and GGUF-compatible environments ### Suggested Use Cases - generating Python, JavaScript, C++, and other programming language snippets - explaining code blocks - rewriting or improving functions - brainstorming implementation strategies - creating scaffolding and prototypes - assisting with debugging and refactoring ## Out-of-Scope Use This model is not guaranteed to be reliable for: - high-stakes legal advice - medical advice - financial decision-making - autonomous execution without review - security-critical production decisions without human verification Users should always validate generated code before deployment. ## Quantization Formats This repository currently includes: - **Q4_K_M** for smaller memory footprint and faster local inference - **Q5_K_M** for improved quality while remaining efficient Choose the quant level based on your hardware budget and quality needs. ## Prompting Notes As a coding-focused conversational model, best results usually come from prompts that are: - specific - structured - explicit about language, framework, or goal - clear about desired output format Example prompt style: > Write a Python function that parses a CSV file, removes duplicate rows by email, and saves the cleaned result. Include error handling and comments. ## Limitations Like other language models, this model may: - hallucinate APIs or library behavior - generate insecure or inefficient code - make reasoning mistakes - produce outdated patterns - require prompt iteration for best results Human review is strongly recommended, especially for production code. ## License This repository uses a **custom WithIn Us AI license approach**. - The base model may be subject to its original upstream license and terms. - Third-party datasets remain the property of their respective creators / licensors. - WithIn Us AI claims authorship of the fine-tuning / merging concept, process, packaging, naming, and release structure for this model distribution. - This repository does **not** claim ownership over third-party datasets or the underlying upstream base model. You can include a `LICENSE` file in this repository with the exact custom terms you want enforced. ## Acknowledgments Special thanks to: - **Qwen** for the base model - all third-party dataset creators listed above - the open-source GGUF / llama.cpp ecosystem - the broader Hugging Face community ## Files Current repository files include: - `WithIn-Us-Coder-4B.Q4_K_M.gguf` - `WithIn-Us-Coder-4B.Q5_K_M.gguf` ## Disclaimer This model may generate incorrect, biased, insecure, or incomplete outputs. Use responsibly, validate important results, and review all generated code before real-world use.