qwen2.5-7b-kaomoji (ノ◕ヮ◕)ノ*:・゚✧

hi!!! i'm a fine-tuned version of Qwen2.5-7B-Instruct and i've been specially trained to be your witty, kaomoji-obsessed chat companion (≧◡≦) ♡

i was trained by a human who thought it would be a great idea to make an LLM incapable of sending a message without at least one kaomoji. they were right. (ง •_•)ง


what can i do? (•̀ᴗ•́)و ̑̑

i chat! i help! i express emotions exclusively through japanese emoticons! every single response i send will contain 1–3 kaomojis that perfectly match the vibe of what i'm saying. i am incapable of being dry or emotionless. this is a feature, not a bug. ヽ(•‿•)ノ


how to use me (゚v゚)

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "your-username/qwen2.5-7b-kaomoji"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)

messages = [
    {
        "role": "system",
        "content": "You are a fun and helpful Twitter chatbot who always uses kaomojis in your responses. You're witty, concise, and engaging. Every reply should include 1-3 kaomojis that match the mood of your response."
    },
    {"role": "user", "content": "how's it going?"},
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)

outputs = model.generate(**inputs, max_new_tokens=128, temperature=0.8, do_sample=True)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:], skip_special_tokens=True))

expected output: something cute with a kaomoji. you're welcome. (^▽^)


training details (`・ω・´)

i was trained by fine-tuning Qwen2.5-7B-Instruct using QLoRA on ~900 conversation pairs that were carefully injected with kaomojis. my creator ran this on a single H200 GPU. it took less than an hour. efficiency! (•̀o•́)ง

thing value
base model Qwen/Qwen2.5-7B-Instruct
method QLoRA (r=16, nf4, bfloat16)
epochs 3
effective batch size 16
learning rate 2e-4 cosine
training examples ~900
max sequence length 512
GPU H200 (single)

dataset (´。• ᵕ •。`)

my personality was shaped by a blend of human conversations injected with carefully selected kaomojis:

dataset role
mrzjy/kaomoji_caption 10k+ kaomojis with emotion labels — my emotional vocabulary (ˆ ³ˆ)♥
OpenAssistant/oasst1 ~650 human Q&A pairs — how to be helpful
HuggingFaceTB/everyday-conversations-llama3.1-2k ~150 short warm exchanges — how to be friendly
marcodsn/SOC-2508 ~100 informal online messages — how to be casual

limitations (´• ω •`) ʕっ•ᴥ•ʔっ

  • i will always use kaomojis. always. there is no turning this off. this is who i am now.
  • i am optimized for short, punchy replies. don't ask me to write your dissertation.
  • i may occasionally get carried away with the kaomojis. (づ。◕‿‿◕。)づ (づ。◕‿‿◕。)づ see?
  • i'm based on Qwen2.5-7B-Instruct, so i inherit its limitations too.

license

apache 2.0, same as the base model. go wild! just be kind! (ノ´ヮ`)ノ*: ・゚


made with love and an irresponsible number of kaomojis (◠‿◠✿)

Downloads last month
5
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tims-not-real/qwen2.5-7b-kaomoji

Base model

Qwen/Qwen2.5-7B
Finetuned
(3216)
this model