Bhagavad Gita Conversational Model

This model is a fine-tuned version of gemma-3-1b-it, trained specifically on the Bhagavad Gita.

Training process:
Stage 1: Standard language modeling on the Bhagavad Gita corpus to capture semantics and style.
Stage 2: Conversation tuning for improved dialogue flow and context-sensitive responses.

Dataset: https://huggingface.co/datasets/nik-55/bhagavad-gita
Training Script: https://huggingface.co/nik-55/bhagavad-gita-gemma-3-1b-it/blob/main/training.ipynb

Note: The model is intended for educational, research only.

How to use

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "nik-55/bhagavad-gita-gemma-3-1b-it"

model_kwargs = dict(
    attn_implementation="eager",
    device_map="auto",
)

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, **model_kwargs)

messages = [
    {
        "role": "user",
        "content": "Who is krishna?",
    },
]

inputs = tokenizer.apply_chat_template(
    messages,
    add_generation_prompt=True,
    tokenize=True,
    return_dict=True,
    return_tensors="pt",
).to("cuda")

outputs = model.generate(**inputs, max_new_tokens=500, do_sample=True)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1] :]))
Downloads last month
3
Safetensors
Model size
1.0B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nik-55/bhagavad-gita-gemma-3-1b-it

Finetuned
(513)
this model

Dataset used to train nik-55/bhagavad-gita-gemma-3-1b-it