YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
π§ Myanmar LLM Training
Fine-tune Qwen2.5-0.5B-Instruct with Myanmar language dataset.
β‘ No License Required!
This model is fully open. No Llama license needed!
π Requirements
- Python 3.8+
- GPU with 6GB+ VRAM
- HuggingFace Account
π Quick Start
1. Install dependencies
pip install -r requirements.txt
2. Login to HuggingFace
huggingface-cli login
3. Run training
python train.py
βοΈ Configuration
| Parameter | Default | Description |
|---|---|---|
| MODEL_NAME | Qwen/Qwen2.5-0.5B-Instruct | Base model (fully open!) |
| num_train_epochs | 3 | Training iterations |
| per_device_train_batch_size | 4 | Batch size |
| gradient_accumulation_steps | 4 | Effective batch = 16 |
| learning_rate | 2e-5 | Learning rate |
π Features
- β Fully open model - ααα―ααΊα ααΊαααα―αα«αααΊα
- β FP16 precision - ααα―ααα―ααΌααΊαα«αααΊα
- β Gradient checkpointing - Memory αα»α½α±αα¬αα«αααΊα
- β Test/Validation evaluation - ααΎα αΊαα―αα―αΆαΈα‘αα½ααΊ α ααΊαΈαααΊαα«αααΊα
π Training Data
Dataset: amkyawdev/AmkyawDev-Dataset
| Split | Samples |
|---|---|
| Train | ~29,100 |
| Validation | ~29,100 |
| Test | ~29,100 |
Note: Each file (train.jsonl, test.jsonl, validation.jsonl) has ~29,100 conversations!
πΎ Output
Trained model saved to ./myanmar-qwen-output/
π€ Upload to HuggingFace
cd myanmar-qwen-output
huggingface-cli upload amkyawdev/my-myanmar-qwen . --repo-type model
π₯οΈ Google Colab
# Install
!pip install transformers datasets torch accelerate
# Login
from huggingface_hub import login
login("YOUR_TOKEN")
# Run
%run train.py
Built by amkyawdev
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support