Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

MetaphoricalCode
/
Dumpling-Qwen2.5-32B-v2-exl3-5bpw-hb6

Text Generation
Transformers
Safetensors
qwen2
conversational
text-generation-inference
5-bit
exl3
Model card Files Files and versions
xet
Community
  • Dumpling-Qwen2.5-32B-v2
    • Method

Quantized using the default exllamav3 (0.0.2) quantization process.

  • Original model: https://huggingface.co/nbeerbower/Dumpling-Qwen2.5-32B-v2
  • exllamav3: https://github.com/turboderp-org/exllamav3

image/png

Dumpling-Qwen2.5-32B-v2

nbeerbower/Rombos-EVAGutenberg-TIES-Qwen2.5-32B finetuned on:

  • nbeerbower/GreatFirewall-DPO
  • nbeerbower/Schule-DPO
  • nbeerbower/Purpura-DPO
  • nbeerbower/Arkhaios-DPO
  • jondurbin/truthy-dpo-v0.1
  • antiven0m/physical-reasoning-dpo
  • flammenai/Date-DPO-NoAsterisks
  • flammenai/Prude-Phi3-DPO
  • Atsunori/HelpSteer2-DPO
  • jondurbin/gutenberg-dpo-v0.1
  • nbeerbower/gutenberg2-dpo
  • nbeerbower/gutenberg-moderne-dpo.

Method

QLoRA ORPO tuned with 8x A100 for 2 epochs. Rank 64 LoRA, 2e-5 learning rate.

Downloads last month
2
Safetensors
Model size
11B params
Tensor type
F16
·
I16
·
Inference Providers NEW
Text Generation
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MetaphoricalCode/Dumpling-Qwen2.5-32B-v2-exl3-5bpw-hb6

Base model

nbeerbower/Rombos-EVAGutenberg-TIES-Qwen2.5-32B
Finetuned
nbeerbower/Dumpling-Qwen2.5-32B-v2
Quantized
(9)
this model

Datasets used to train MetaphoricalCode/Dumpling-Qwen2.5-32B-v2-exl3-5bpw-hb6

jondurbin/gutenberg-dpo-v0.1

Viewer • Updated Jan 12, 2024 • 918 • 411 • 162

jondurbin/truthy-dpo-v0.1

Viewer • Updated Jan 11, 2024 • 1.02k • 381 • 136

nbeerbower/GreatFirewall-DPO

Viewer • Updated Mar 2, 2025 • 492 • 59 • 10
Company
TOS Privacy About Careers
Website
Models Datasets Spaces Pricing Docs