--- title: Qwen API emoji: 🤖 colorFrom: blue colorTo: purple sdk: docker app_file: app.py pinned: false license: apache-2.0 tags: - qwen - uncensored - llama-cpp - gguf suggested_hardware: a10g-small --- # Qwen3.5-9B Uncensored API Interface API interface for [HauhauCS/Qwen3.5-9B-Uncensored-HauhauCS-Aggressive](https://huggingface.co/HauhauCS/Qwen3.5-9B-Uncensored-HauhauCS-Aggressive). ## Features - 9B parameters with 262K context window - Fully uncensored (0/465 refusals) - Multimodal capable (text, image, video) - Supports 201 languages - Q4_K_M quantization via llama.cpp ## API Usage ### Python ```python from gradio_client import Client client = Client("Ngixdev/qwen-api") result = client.predict( prompt="Your question here", system_prompt="You are a helpful assistant", temperature=0.7, top_p=0.8, max_tokens=1024, api_name="/api_generate" ) print(result) ``` ### cURL ```bash curl -X POST https://ngixdev-qwen-api.hf.space/api/api_generate \ -H "Content-Type: application/json" \ -d '{ "data": [ "Your question here", "You are a helpful assistant", 0.7, 0.8, 1024 ] }' ``` ## Parameters | Parameter | Type | Default | Description | |-----------|------|---------|-------------| | prompt | string | required | User prompt/question | | system_prompt | string | "" | System instruction | | temperature | float | 0.7 | Sampling temperature (0.0-2.0) | | top_p | float | 0.8 | Nucleus sampling (0.0-1.0) | | max_tokens | int | 1024 | Maximum tokens to generate |