A full fine-tune of unsloth/gemma-3-270m-it on the kth8/cron-20000x dataset.

Usage example

System prompt

You are a helpful assistant that convert cron expression into human-readable description.

User prompt

33-46 */9 9,13,21,22 * 1-4

Model Details

  • Base Model: unsloth/gemma-3-270m-it
  • Parameter Count: 268098176
  • Training Method: Full Fine-Tune (FFT) - All parameters updated.
  • Precision: torch.bfloat16

Hardware

  • GPU: NVIDIA L4

Training stats

  • Global step: 2307
  • Training runtime: 5201.0361
  • Average training loss: 0.02166572387309018
  • Final validation loss: 0.0001278361160075292
  • Epoch: 1.0

Framework versions

  • Unsloth: 2026.3.4
  • TRL: 0.22.2
  • Transformers: 4.56.2
  • Pytorch: 2.10.0+cu128
  • Datasets: 4.3.0
  • Tokenizers: 0.22.2

License

This model is released under the Gemma license. See the Gemma Terms of Use for details.

Downloads last month
5
Safetensors
Model size
0.3B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for kth8/gemma-3-270m-it-Cron-Converter

Finetuned
(392)
this model
Quantizations
1 model

Dataset used to train kth8/gemma-3-270m-it-Cron-Converter