--- license: mit base_model: google/gemma-3-270m-it tags: - causal-reasoning - semantic-loss - fine-tuned - d-separation language: - en --- # Gemma D-Separation Semantic V2 Fine-tuned Gemma 270M-IT for d-separation causal reasoning using semantic loss with dynamic lambda scheduling. ## Performance - Standard accuracy: 68.6% - Adversarial accuracy: 67.8% - F1 score: 25.0% (vs 7.6% collapsed baseline) ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained("ludwigw/gemma-dseparation-semantic-v2") tokenizer = AutoTokenizer.from_pretrained("ludwigw/gemma-dseparation-semantic-v2") ``` ## Citation ```bibtex @article{deshmukh2026semantic, title={On Semantic Loss Fine-Tuning Approach for Preventing Model Collapse in Causal Reasoning}, author={Deshmukh, Pratik and Gupta, Atirek}, journal={arXiv preprint arXiv:2605.05438}, year={2026} } ```