Compressed Model: MilyaShams/Qwen3-1.7B-SparseGPT_2_4_bs64

This model was compressed using the llmcompressor framework.

Compression Details

  • Base Model: Qwen/Qwen3-1.7B
  • Experiment Name: SparseGPT_2_4_bs64
  • Recipe / Modifiers Applied:
index=None group=None start=None end=None update=None initialized_=True finalized_=True started_=True ended_=True sparsity=0.5 sparsity_profile=None mask_structure='2:4' owl_m=None owl_lmbda=None sequential_update=False sequential_targets=['Qwen3DecoderLayer'] targets=['Linear'] ignore=[] block_size=64 dampening_frac=0.05 preserve_sparsity_mask=False offload_hessians=False

Note: This model card was automatically generated. All structural modifiers and parameters used during compression are logged above.

Downloads last month
113
Safetensors
Model size
1B params
Tensor type
I64
F16
U8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for MilyaShams/Qwen3-1.7B-SparseGPT_2_4_bs64

Finetuned
Qwen/Qwen3-1.7B
Quantized
(254)
this model