Compressed Model: MilyaShams/Qwen3-1.7B-AutoRound_W8A8
This model was compressed using the llmcompressor framework.
Compression Details
- Base Model: Qwen/Qwen3-1.7B
- Experiment Name: AutoRound_W8A8
- Recipe / Modifiers Applied:
config_groups=None targets=['Linear'] ignore=[] scheme='W8A8' kv_cache_scheme=None weight_observer=None input_observer=None output_observer=None observer=None bypass_divisibility_checks=False index=None group=None start=None end=None update=None initialized_=True finalized_=True started_=True ended_=True sequential_targets=['Qwen3DecoderLayer'] iters=200 enable_torch_compile=True batch_size=2 lr=None device_ids='0,1'
Note: This model card was automatically generated. All structural modifiers and parameters used during compression are logged above.
- Downloads last month
- 137
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support