VisMem RAG Fine-tuned: gemma-3-270m-vismem-rag-0316-1016

Inference

import requests
from vismem_core import VisMem
from sentence_transformers import SentenceTransformer
from transformers import pipeline
# 1. Load VisMem
data = requests.get(
    "https://huggingface.co/datasets/broadfield-dev/gemma-3-270m-vismem-kb-0316-1016/resolve/main/vismem.png",
    headers={"Authorization":"Bearer <HF_TOKEN>"}).content
mem  = VisMem.from_png_bytes(data)
emb  = SentenceTransformer('all-MiniLM-L6-v2')
# 2. RAG query
q_vec   = emb.encode([your_question])[0]
results = mem.search(q_vec, k=3)
context = "\n---\n".join(results)
# 3. Prompt
system = (
    "You are a helpful AI Assistant with visual memory.\n"
    "### RAG MEMORY (Vector Database):\n"
    "[Uploaded Doc]: None\n"
    f"[Knowledge Base]: {context}\n"
    "### EPISODIC MEMORY (Past Chat):\n"
    "[History]: None"
)
pipe = pipeline("text-generation", model="broadfield-dev/gemma-3-270m-vismem-rag-0316-1016")
print(pipe([
    {"role":"system","content":system},
    {"role":"user","content":your_question}
], max_new_tokens=200))

Config

{ "dataset_name": "nohurry/Opus-4.6-Reasoning-3000x-filtered", "rag_columns": [ "problem", "thinking" ], "question_col": "problem", "answer_col": "solution", "split": "train", "data_config": null, "total_kb_docs": 2326, "vismem_dim": 384, "vismem_width": 8192, "vismem_height": 8192, "kb_repo": "broadfield-dev/gemma-3-270m-vismem-kb-0316-1016" }

Downloads last month
13
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support