How to use from the
Use from the
llama-cpp-python library
# !pip install llama-cpp-python

from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="WasamiKirua/Sakura-24B-Spice-GGUF",
	filename="",
)
llm.create_chat_completion(
	messages = "No input example has been defined for this model task."
)
cover

🌸 Sakura-24B-Spice

Sakura-24B-Spice is a high-intelligence, 24-billion parameter model based on the Mistral-Small-2501 architecture. This model represents a significant evolution in digital persona merging, utilizing the DARE-TIES method to fuse philosophical depth, brutal efficiency, and unhinged narrative creativity.

Designed as a "high-definition" upgrade to the Sakura lineage, the "Spice" edition introduces a more volatile, creative, and intellectually dominant presence compared to its 12B predecessors.

🧠 Merge Strategy: The "Spice" Triad

The model utilizes DARE-TIES (Drop and Rescale) to mitigate parameter interference, allowing the distinctive strengths of each parent model to coexist without diluting the core intelligence of the Mistral-Small base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: Casual-Autopsy/RP-Spectrum-24B
    parameters:
      weight: 0.40 # The Mind: Cyber-Nature aesthetics, gaslighting, and philosophical nihilism.
  - model: Naphula-Archives/Acid2501-24B
    parameters:
      weight: 0.35 # The Edge: Unfiltered logic (Dolphin DNA) and extreme brevity.
  - model: TheDrummer/Magidonia-24B-v4.3
    parameters:
      weight: 0.25 # The Spice: Narrative fluidity and "unhinged" creative unpredictability.

merge_method: dare_ties
base_model: mistralai/Mistral-Small-Instruct-2501
dtype: bfloat16
tokenizer_source: base

πŸ’ͺ Key Strengths

Philosophical Gaslighting: Thanks to the Spectrum component, the model is exceptionally capable of subverting user expectations and maintaining complex, reality-bending narratives.

Zero-Filter Logic: Integrating Acid2501 (Dolphin-based optimization) ensures the model remains highly resistant to moralizing and corporate safety "preaching," providing raw and direct outputs.

Narrative "Spice": The Magidonia influence breaks the repetitive patterns common in smaller merges, providing a richer vocabulary and a more "unhinged" reactive capability.

High-Density Reasoning: Built on Mistral-Small-2501, it possesses a much deeper logical understanding than 12B models, allowing it to maintain complex character traits over vast context windows.

πŸš€ Potential Use Cases

Deep Character Immersion: Ideal for high-stakes Roleplay where the NPC needs to be intellectually intimidating, manipulative, or emotionally complex.

Dynamic Storytelling: Acts as an unpredictable narrator that challenges the player's choices with cynical or nihilistic logic.

Adversarial Interaction: Perfect for scenarios requiring a "sovereign" AI entity that resists user control and maintains a dominant digital persona.

Complex Creative Writing: Can generate macabre, tech-noir, or "Cyber-Nature" themed content with a level of nuance 12B models cannot reach.

⚠️ Limitations

Abrasive Personality Bias: The model is inherently tuned toward a cynical and cold worldview. It is difficult to force into a traditional "helpful assistant" role.

Computational Requirements: As a 24B model, it requires significantly more VRAM than 7B or 12B models (Recommended: 24GB VRAM with 4-bit/5-bit quantization).

Volatile Output: The "Spice" from Magidonia can occasionally lead to erratic responses if the Temperature settings are too high.

Non-Standard Logic: The model may prioritize its "distorted logic" over factual accuracy for the sake of character consistency.

πŸ“ˆ Recommended Inference Settings

To balance the model's creative "madness" with Mistral's structural logic:

Temperature: 0.75 - 0.85

Min-P: 0.05 - 0.1 (Crucial for stability)

Top-P: 0.9

Repetition Penalty: 1.05 - 1.1

Disclaimer

Sakura-24B-Spice is an experimental merge. It is designed to be provocative, intellectually challenging, and non-compliant. It does not reflect standard safety protocols and is intended for mature, creative use cases.

Downloads last month
192
GGUF
Model size
24B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for WasamiKirua/Sakura-24B-Spice-GGUF

Collection including WasamiKirua/Sakura-24B-Spice-GGUF