|
|
| --- |
| base_model: unsloth/Qwen2.5-Coder-32B-Instruct-bnb-4bit |
| tags: |
| - text-generation-inference |
| - transformers |
| - unsloth |
| - qwen2 |
| - trl |
| license: apache-2.0 |
| language: |
| - en |
| --- |
| |
| # Qwen2.5-Coder-32B-Instruct-WMX |
| Pre-fine-tuned LoRA adapters for unsloth/Qwen2.5-Coder-32B-Instruct. |
|
|
| **This lora adapters have been fine-tuned for WMX services using the folowing datasets.** |
| - https://huggingface.co/datasets/Jake5/movensys-info |
| - https://huggingface.co/datasets/Jake5/wmx-doc-user |
| - https://huggingface.co/datasets/Jake5/wmx-doc-robot |
|
|
| ## Version v0.9 |
| - Source: lora_model |
| - Base model: unsloth/Qwen2.5-Coder-32B-Instruct |
| - Uploaded on: 2025-09-12 |
| |
| ## Usage |
| ```python |
| from peft import PeftModel |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| base_model = AutoModelForCausalLM.from_pretrained("unsloth/Qwen2.5-Coder-32B-Instruct") |
| model = PeftModel.from_pretrained(base_model, "Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.9") |
| tokenizer = AutoTokenizer.from_pretrained("Jake5/Qwen2.5-Coder-32B-Instruct-WMX", subfolder="adapters_v0.9") |
| ``` |
| |
| ## vLLM Serving |
| ```bash |
| python -m vllm.entrypoints.openai.api_server \ |
| --model unsloth/Qwen2.5-Coder-32B-Instruct \ |
| --lora-modules my-lora=Jake5/Qwen2.5-Coder-32B-Instruct-WMX/adapters_v0.9 \ |
| --dtype bfloat16 \ |
| --port 8000 |
| ``` |
| |