anime-translator-v6 (Beta)
This model is in Beta. It was trained and benchmarked on a single show (Jujutsu Kaisen S01E01 vs. Crunchyroll DE ground truth). Results may vary on other series. Feedback and corrections welcome โ open an issue or reach out via Sublarr.
A Gemma-3-12B fine-tuned ENโDE subtitle translation model specialized for anime dialogue. Designed to run locally via Ollama, integrated as a community backend in Sublarr.
Why this model?
General-purpose LLMs add meta-commentary ("Here is the translation:") that breaks subtitle formatting. Dedicated document translators (e.g. Hunyuan-MT-7B) hallucinate on short 2โ5 word anime lines. This model is fine-tuned specifically on anime subtitle pairs โ clean, concise output at 7 GB.
Benchmark (JJK S01E01 vs. Crunchyroll DE, 30 aligned pairs)
| Model | BLEU-1 | BLEU-2 | Length-Ratio | Issues |
|---|---|---|---|---|
| anime-translator-v6 | 0.281 | 0.111 | 1.02 | 0 meta-commentary |
| qwen2.5:14b | 0.264 | 0.125 | 0.96 | Adds "Hier ist die รbersetzung:" |
| hunyuan-mt-7b | 0.141 | โ | 12.76 | Verbose, hallucinations on short lines |
V6 wins on BLEU-1 (most relevant for subtitle length) and produces the cleanest output.
Training Details
| Parameter | Value |
|---|---|
| Base model | google/gemma-3-12b-it |
| Training data | OPUS OpenSubtitles v2018 EN-DE (~75k filtered pairs) |
| LoRA Rank | 32 / Alpha 64 |
| Epochs | 2 |
| Learning rate | 1e-4 |
| Max seq len | 256 |
| Hardware | RTX 5090 (32 GB VRAM, ~80 min) |
| Quantization | Q4_K_M GGUF |
Key design decision: Uses apply_chat_template consistently for both training and inference.
This format consistency is the core fix over V1โV5 which used seq2seq bases with mismatched formats.
Quick Start with Ollama
# Option 1: Direct pull from HuggingFace (Ollama โฅ 0.3)
ollama pull hf.co/sublarr/anime-translator-v6-GGUF:Q4_K_M
# Option 2: Manual create with Modelfile
git clone https://huggingface.co/sublarr/anime-translator-v6-GGUF
ollama create anime-translator-v6 -f anime-translator-v6-GGUF/Modelfile
Sublarr Integration
This model is available as a one-click community backend in Sublarr:
- Open Sublarr โ Settings โ Translation Backends
- Select "Community Models" โ "Anime Translator V6"
- Click "Install" โ Ollama pulls the model automatically
- Set as active backend โ done
# Or set manually in sublarr.env:
SUBLARR_OLLAMA_MODEL=anime-translator-v6
Prompt Format (Gemma-3 Chat Template)
<start_of_turn>user
Translate to German:
{english_subtitle_text}<end_of_turn>
<start_of_turn>model
Inference Parameters
temperature=0.3
top_p=0.9
top_k=40
repeat_penalty=1.1
num_predict=200
Limitations
- Beta quality: Tested on one show, may degrade on other genres/styles
- ENโDE only: Not designed for other language pairs
- Short lines only: Optimized for subtitle-length text (< 20 words), not document translation
- Anime-specific: May underperform on live-action, documentary, or neutral-register content
Roadmap
- Broader test set (5+ shows, different studios)
- Upload LoRA adapter weights separately (for further fine-tuning)
- V7: Explore NLLB-200 fine-tuning as seq2seq alternative
License
MIT โ model weights derived from google/gemma-3-12b-it (Gemma Terms of Use apply).
Training data: OPUS OpenSubtitles v2018 (CC BY 4.0).
- Downloads last month
- 13
4-bit