Qwen3-4B-Nymphaea-RP

A fine-tune of Qwen3-4B-Instruct-2507 for roleplay and creative writing.

Suitable for mobile roleplay: tested on Nothing Phone 2 in Q4_K_M quantization (7-8 t/s)

The SillyTavern preset is available here. For custom presets, please use the ChatML instruct template.

Chat Example

Tested at Q8_0 quantization.

SillyTavern Screenshot

Training Notes

Trained on the latest iteration of my Darkmere dataset. This version features expanded genre variety, built upon a mix of manually curated synthetics and human-written stories.

The base weights are abliterated via Heretic prior to fine-tuning, so this fine-tune is quite uncensored.

Training Specs

Method:

  • Training Method: DoRA (Weight-Decomposed LoRA)
  • Target Modules all-linear
  • LoRA Rank: 32
  • LoRA Alpha: 32
  • LoRA Dropout: 0.05

Hyperparameters:

  • Batch Size: 2 (Per-device)
  • Gradient Accumulation: 2
  • Epochs: 2
  • Learning Rate: 1e-4
  • Optimizer: adamw_torch_fused
  • LR Scheduler: cosine
  • Noise Level: neftune_noise_alpha=5

Special Thanks

This fine-tune wouldn't be possible without the incredible work of the community:

Downloads last month
597
Safetensors
Model size
4B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for 0xA50C1A1/Qwen3-4B-Nymphaea-RP

Finetuned
(1)
this model
Quantizations
2 models

Collection including 0xA50C1A1/Qwen3-4B-Nymphaea-RP