Qwen3-0.6B-grok

Fine-tuned version of Qwen/Qwen3-0.6B for Grok parser pattern generation.

Description

Given a raw log line, generates the appropriate parse_grok!() pattern that extracts structured fields.

Training Details

  • Base Model: Qwen/Qwen3-0.6B
  • Method: LoRA (r=64, alpha=64) + 4-bit quantization
  • Training Samples: 646973
  • Epochs: 1
  • Learning Rate: 0.0002
  • Final Training Loss: 0.3573894602136961
  • Training Time: 0.0 minutes
  • Hardware: NVIDIA H100 80GB
  • Framework: Unsloth + TRL

Usage

Model card will be updated with benchmark results after evaluation.

Downloads last month
-
Safetensors
Model size
0.6B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for omeryentur/Qwen3-0.6B-grok

Finetuned
Qwen/Qwen3-0.6B
Finetuned
(800)
this model