--- language: code tags: - code-generation - python - fine-tuned - qlora base_model: Qwen/Qwen2.5-Coder-0.5B-Instruct datasets: - iamtarun/python_code_instructions_18k_alpaca license: mit --- # Qwen2.5-Coder-0.5B Python Fine-tuned Fine-tuned version of [Qwen/Qwen2.5-Coder-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-0.5B-Instruct) for Python code generation. ## Model Details - **Base Model**: Qwen/Qwen2.5-Coder-0.5B-Instruct - **Fine-tuning Method**: QLoRA (4-bit quantization + LoRA adapters) - **Dataset**: iamtarun/python_code_instructions_18k_alpaca - **Task**: Python code generation from natural language instructions ## Training Details - **Training Samples**: 16000 - **Validation Samples**: 1000 - **Epochs**: 3 - **Training Time**: N/A - **Final Loss**: N/A ## Performance - **Syntax Validity**: 95.2% - **Pass@1**: 54.4% - **Verbosity Reduction**: 95% ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("KpRT/qwen-python-finetuned") tokenizer = AutoTokenizer.from_pretrained("KpRT/qwen-python-finetuned") prompt = "Write a function to reverse a string" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=256) code = tokenizer.decode(outputs[0], skip_special_tokens=True) print(code) ``` ## Citation If you use this model, please cite: ```bibtex @misc{qwen-python-finetuned, author = {K R T}, title = {Qwen2.5-Coder Python Fine-tuned}, year = {2026}, publisher = {HuggingFace}, url = {https://huggingface.co/KpRT/qwen-python-finetuned} } ```