Gemma3 CPT
Collection
9 items • Updated
This model is the 2nd-epoch Stage 2 Turkish Continued Pretraining (CPT) variant of Gemma-3-1B.
Unlike the initial epoch stages that started from google/gemma-3-1b-pt,
this model is initialized from the previous stage of epoch-2 checkpoint:
canbingol/gemma3_1B_base-tr-cpt-2nd_epoch_stage1Stage 2 of the second epoch continues domain adaptation through the next shard in the sequential training pipeline.
The model was trained for 1 epoch on samples 50,000 to 100,000 of the Turkish web corpus.
Conceptually:
google/gemma-3-1b-ptThis represents sequential CPT across disjoint data shards, repeated for a second epoch.
canbingol/vngrs-web-corpus-200kcanbingol/gemma3_1B_base-tr-cpt-2nd_epoch_stage1import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "canbingol/gemma3_1B_base-tr-cpt-2nd_epoch_stage2"
device = "cuda" if torch.cuda.is_available() else "cpu"
model = AutoModelForCausalLM.from_pretrained(model_id)
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = model.to(device)
prompt = "bundan böyle"
inputs = tokenizer(prompt, return_tensors="pt").to(device)
outputs = model.generate(
**inputs,
max_new_tokens=50,
do_sample=True,
temperature=0.8,
top_p=0.9
)
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(generated_text)
Base model
google/gemma-3-1b-pt