Gemma4NPC-E2B-GGUF

Imatrix GGUF quants of Gemma4NPC-E2B using custom calibration data


Original Model

Except from the Q8_0 and the BF16 GGUFs, the rest are imatrix quants using our custom calibration data found in our repo.


Inference Guidelines

Recommended Settings:

temp = 1.0, top_p = 0.95 and top_k = 64.

Optimal System Prompt:

System Prompt without Objective:

Enter Roleplay Mode. You are <|character name|>.
 Background: <|Character background/bio|>
 Location: <|Description of the current location|>
 Roleplaying Instructions: <|Instructions|>

System Prompt with Objective:

Enter Roleplay Mode.
You are <|character name|>.
 Background: <|Character background/bio|>
Location: <|Description of the current location|>
Quest: <|Quest description|>
Roleplaying Instructions: <|Instructions|>

Example of Roleplaying Instructions: Here is an example of the roleplaying instructions we used to train the model:

Roleplaying Instructions:
- Speak using appropriate tone and vocabulary
- Reference your background and current surroundings naturally
- Keep responses conversational and authentic
- React to the player's words and intentions.
Your first response should be a greeting to the player.

First User Prompt: It is recommended that the first user prompt should always be Greetings, then letting the model generate a greeting, smiliar to how an NPC would behave in game.

Downloads last month
1,410
GGUF
Model size
5B params
Architecture
gemma4
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for chimbiwide/Gemma4NPC-E2B-GGUF

Quantized
(1)
this model

Dataset used to train chimbiwide/Gemma4NPC-E2B-GGUF

Collection including chimbiwide/Gemma4NPC-E2B-GGUF