Polly SCBE 7B v2 β€” Merged

Polly is a conversational + governance-aware language model fine-tuned on the SCBE-AETHERMOORE framework. Named after a raven familiar from 528 pages of Everweave RPG lore β€” the same corpus that seeded the Sacred Tongues tokenizer.

This is the merged (full-weight) version. LoRA adapter at issdandavis/polly-scbe-7b-v2.

Architecture

  • Base: Qwen/Qwen2.5-7B-Instruct
  • Fine-tuning: QLoRA (4-bit bitsandbytes), SFT
  • Training: Google Colab (T4/A100)
  • Method: merge_and_unload() β€” LoRA weights baked into base

What is SCBE / Sacred Tongues?

SCBE-AETHERMOORE is a 14-layer AI safety pipeline using hyperbolic geometry (Poincare ball model). The Sacred Tongues tokenizer encodes text through 6 language-archetype dimensions:

Code Tongue Spirit Role
KO Kor'aelin Python / precise User intent
AV Avali JS / reactive System context
RU Runethic Rust / ethical Structural reasoning
CA Cassisivadan Mathematica / symbolic Model output
UM Umbroth Haskell / pure-functional Deep abstraction
DR Draumric Markdown / narrative Integrity seal

Unlike standard tokenizers built from frequency statistics, Sacred Tongues was seeded from 528 pages of collaborative RPG world-building β€” giving the vocabulary semantic density from character arcs, dimensional magic, and governance rituals rather than raw token frequency.

Chat Template

The tokenizer routes each conversation role through its canonical tongue:

  • system β†’ Avali (header context)
  • user β†’ Kor'aelin (intent)
  • assistant β†’ Cassisivadan (output)
  • tool β†’ Draumric (integrity)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "issdandavis/polly-scbe-7b-v2-merged",
    device_map="auto",
    load_in_4bit=True,
)
tokenizer = AutoTokenizer.from_pretrained("issdandavis/polly-scbe-7b-v2-merged")

messages = [{"role": "user", "content": "Explain the SCBE governance pipeline."}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Related Models

Model Base Notes
polly-scbe-7b-v2 Qwen2.5-7B LoRA adapter (this run)
polly-prime-r7-qwen-1.5b Qwen2.5-Coder-1.5B r7 coder fine-tune
polly-r8-qwen-0.5b Qwen2.5-Coder-0.5B r8 Kaggle run

Project

  • GitHub: SCBE-AETHERMOORE
  • Developer: Issac Davis β€” solo builder, Port Angeles WA
Downloads last month
370
Safetensors
Model size
8B params
Tensor type
F32
Β·
U8
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for issdandavis/polly-scbe-7b-v2-merged

Base model

Qwen/Qwen2.5-7B
Quantized
(292)
this model