Spaces:
Sleeping
Sleeping
Upload app.py
Browse files
app.py
CHANGED
|
@@ -1,43 +1,49 @@
|
|
| 1 |
-
|
| 2 |
-
import gradio as gr
|
| 3 |
import os
|
| 4 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
# Model options
|
| 6 |
MODELS = [
|
| 7 |
-
# Small (fast, <3B)
|
| 8 |
'Qwen/Qwen2.5-0.5B-Instruct',
|
| 9 |
-
'Qwen/Qwen2.5-1.5B-Instruct',
|
| 10 |
-
'microsoft/Phi-3-mini-128k-instruct',
|
| 11 |
-
'google/gemma-2b-it',
|
| 12 |
-
'meta-llama/Llama-3.2-1B-Instruct',
|
| 13 |
-
# Medium (7-13B)
|
| 14 |
'Qwen/Qwen2.5-7B-Instruct',
|
|
|
|
| 15 |
'meta-llama/Llama-3.2-3B-Instruct',
|
|
|
|
|
|
|
| 16 |
'mistralai/Mistral-7B-Instruct-v0.3',
|
| 17 |
-
# Large (needs more GPU)
|
| 18 |
-
'Qwen/Qwen2.5-14B-Instruct',
|
| 19 |
-
'mistralai/Mixtral-8x7B-Instruct-v0.1',
|
| 20 |
]
|
| 21 |
|
| 22 |
-
|
| 23 |
-
METHODS = ['SFT', 'DPO', 'GRPO', 'LoRA']
|
| 24 |
|
| 25 |
-
# Public datasets
|
| 26 |
DATASETS = [
|
| 27 |
'HuggingFaceH4/ultrachat_200k',
|
| 28 |
'openai/gsm8k',
|
| 29 |
'meta-math/MATH',
|
| 30 |
'anthropic/hh-rlhf',
|
| 31 |
-
'stanfordnlp/SHP',
|
| 32 |
]
|
| 33 |
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
|
|
|
| 40 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 41 |
from trl import SFTTrainer, SFTConfig
|
| 42 |
from datasets import load_dataset
|
| 43 |
|
|
@@ -67,14 +73,8 @@ trainer = SFTTrainer(
|
|
| 67 |
trainer.train()
|
| 68 |
trainer.push_to_hub()
|
| 69 |
'''
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
def generate_dpo_script(model, dataset, epochs, lr, batch_size, output_name):
|
| 73 |
-
template = f'''#!/usr/bin/env python3
|
| 74 |
-
# DPO Training Script
|
| 75 |
-
# Model: {model}
|
| 76 |
-
# Dataset: {dataset}
|
| 77 |
-
|
| 78 |
from trl import DPOTrainer, DPOConfig
|
| 79 |
from datasets import load_dataset
|
| 80 |
|
|
@@ -84,7 +84,7 @@ training_args = DPOConfig(
|
|
| 84 |
learning_rate={lr},
|
| 85 |
num_train_epochs={epochs},
|
| 86 |
per_device_train_batch_size={batch_size},
|
| 87 |
-
max_seq_length=
|
| 88 |
bf16=True,
|
| 89 |
output_dir='./{output_name}-output',
|
| 90 |
push_to_hub=True,
|
|
@@ -100,14 +100,8 @@ trainer = DPOTrainer(
|
|
| 100 |
trainer.train()
|
| 101 |
trainer.push_to_hub()
|
| 102 |
'''
|
| 103 |
-
|
| 104 |
-
|
| 105 |
-
def generate_lora_script(model, dataset, epochs, lr, batch_size, output_name):
|
| 106 |
-
template = f'''#!/usr/bin/env python3
|
| 107 |
-
# LoRA Fine-tuning Script
|
| 108 |
-
# Model: {model}
|
| 109 |
-
# Dataset: {dataset}
|
| 110 |
-
|
| 111 |
from trl import SFTTrainer, SFTConfig
|
| 112 |
from peft import LoraConfig
|
| 113 |
from datasets import load_dataset
|
|
@@ -122,7 +116,7 @@ peft_config = LoraConfig(
|
|
| 122 |
)
|
| 123 |
|
| 124 |
training_args = SFTConfig(
|
| 125 |
-
learning_rate={lr * 10},
|
| 126 |
num_train_epochs={epochs},
|
| 127 |
per_device_train_batch_size={batch_size},
|
| 128 |
max_seq_length=2048,
|
|
@@ -142,69 +136,27 @@ trainer = SFTTrainer(
|
|
| 142 |
trainer.train()
|
| 143 |
trainer.push_to_hub()
|
| 144 |
'''
|
| 145 |
-
|
| 146 |
-
|
| 147 |
-
|
| 148 |
-
|
| 149 |
-
return generate_dpo_script(model, dataset, epochs, lr, batch_size, output_name)
|
| 150 |
-
elif method == 'LoRA':
|
| 151 |
-
return generate_lora_script(model, dataset, epochs, lr, batch_size, output_name)
|
| 152 |
-
else: # SFT, GRPO
|
| 153 |
-
return generate_training_script(model, method, dataset, epochs, lr, batch_size, max_length, output_name)
|
| 154 |
-
|
| 155 |
-
def get_hardware_requirement(model):
|
| 156 |
if '0.5B' in model or '1B' in model:
|
| 157 |
-
|
| 158 |
elif '3B' in model or '7B' in model:
|
| 159 |
-
|
| 160 |
else:
|
| 161 |
-
|
| 162 |
-
|
| 163 |
-
# Gradio UI
|
| 164 |
-
with gr.Blocks(title='ML Model Trainer', theme=gr.themes.Soft()) as demo:
|
| 165 |
-
gr.Markdown('# 🤖 ML Model Trainer')
|
| 166 |
-
gr.Markdown('Generate training scripts for SFT, DPO, LoRA fine-tuning')
|
| 167 |
-
|
| 168 |
-
with gr.Row():
|
| 169 |
-
with gr.Column():
|
| 170 |
-
model = gr.Dropdown(MODELS, label='Model', value='Qwen/Qwen2.5-0.5B-Instruct')
|
| 171 |
-
method = gr.Dropdown(METHODS, label='Training Method', value='SFT')
|
| 172 |
-
dataset = gr.Dropdown(DATASETS, label='Dataset', value='HuggingFaceH4/ultrachat_200k')
|
| 173 |
-
output_name = gr.Textbox(label='Output Model Name', value='my-finetuned-model')
|
| 174 |
-
|
| 175 |
-
with gr.Column():
|
| 176 |
-
epochs = gr.Slider(1, 10, value=3, step=1, label='Epochs')
|
| 177 |
-
lr = gr.Number(label='Learning Rate', value=2e-5)
|
| 178 |
-
batch_size = gr.Slider(1, 16, value=4, step=1, label='Batch Size')
|
| 179 |
-
max_length = gr.Slider(256, 8192, value=4096, step=256, label='Max Sequence Length')
|
| 180 |
-
|
| 181 |
-
generate_btn = gr.Button('🔧 Generate Training Script', variant='primary')
|
| 182 |
|
| 183 |
-
|
| 184 |
-
|
| 185 |
-
|
| 186 |
-
|
| 187 |
-
|
| 188 |
-
|
| 189 |
-
|
| 190 |
-
|
| 191 |
-
|
| 192 |
-
|
| 193 |
-
|
| 194 |
-
|
| 195 |
-
|
| 196 |
-
)
|
| 197 |
-
|
| 198 |
-
gr.Markdown('---')
|
| 199 |
-
gr.Markdown('### 📋 How to Use')
|
| 200 |
-
gr.Markdown('''
|
| 201 |
-
1. Configure your training parameters above
|
| 202 |
-
2. Click **Generate Training Script**
|
| 203 |
-
3. Copy the script to a file (e.g., `train.py`)
|
| 204 |
-
4. Install dependencies: `pip install transformers trl torch datasets accelerate peft`
|
| 205 |
-
5. Run: `python train.py`
|
| 206 |
-
|
| 207 |
-
**Note:** You'll need Hugging Face Pro or compute credits for cloud training.
|
| 208 |
-
''')
|
| 209 |
-
|
| 210 |
-
demo.launch()
|
|
|
|
| 1 |
+
import streamlit as st
|
|
|
|
| 2 |
import os
|
| 3 |
|
| 4 |
+
st.set_page_config(page_title='ML Model Trainer', page_icon='🤖')
|
| 5 |
+
|
| 6 |
+
st.title('🤖 ML Model Trainer')
|
| 7 |
+
st.markdown('Generate training scripts for SFT, DPO, LoRA fine-tuning')
|
| 8 |
+
|
| 9 |
# Model options
|
| 10 |
MODELS = [
|
|
|
|
| 11 |
'Qwen/Qwen2.5-0.5B-Instruct',
|
| 12 |
+
'Qwen/Qwen2.5-1.5B-Instruct',
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
'Qwen/Qwen2.5-7B-Instruct',
|
| 14 |
+
'meta-llama/Llama-3.2-1B-Instruct',
|
| 15 |
'meta-llama/Llama-3.2-3B-Instruct',
|
| 16 |
+
'microsoft/Phi-3-mini-128k-instruct',
|
| 17 |
+
'google/gemma-2b-it',
|
| 18 |
'mistralai/Mistral-7B-Instruct-v0.3',
|
|
|
|
|
|
|
|
|
|
| 19 |
]
|
| 20 |
|
| 21 |
+
METHODS = ['SFT', 'DPO', 'LoRA']
|
|
|
|
| 22 |
|
|
|
|
| 23 |
DATASETS = [
|
| 24 |
'HuggingFaceH4/ultrachat_200k',
|
| 25 |
'openai/gsm8k',
|
| 26 |
'meta-math/MATH',
|
| 27 |
'anthropic/hh-rlhf',
|
|
|
|
| 28 |
]
|
| 29 |
|
| 30 |
+
col1, col2 = st.columns(2)
|
| 31 |
+
|
| 32 |
+
with col1:
|
| 33 |
+
model = st.selectbox('Model', MODELS)
|
| 34 |
+
method = st.selectbox('Training Method', METHODS)
|
| 35 |
+
dataset = st.selectbox('Dataset', DATASETS)
|
| 36 |
+
output_name = st.text_input('Output Model Name', 'my-finetuned-model')
|
| 37 |
|
| 38 |
+
with col2:
|
| 39 |
+
epochs = st.slider('Epochs', 1, 10, 3)
|
| 40 |
+
lr = st.number_input('Learning Rate', value=2e-5, format='%.0e')
|
| 41 |
+
batch_size = st.slider('Batch Size', 1, 16, 4)
|
| 42 |
+
max_length = st.slider('Max Sequence Length', 256, 8192, 4096, 256)
|
| 43 |
+
|
| 44 |
+
if st.button('🔧 Generate Training Script', type='primary'):
|
| 45 |
+
if method == 'SFT':
|
| 46 |
+
script = f'''#!/usr/bin/env python3
|
| 47 |
from trl import SFTTrainer, SFTConfig
|
| 48 |
from datasets import load_dataset
|
| 49 |
|
|
|
|
| 73 |
trainer.train()
|
| 74 |
trainer.push_to_hub()
|
| 75 |
'''
|
| 76 |
+
elif method == 'DPO':
|
| 77 |
+
script = f'''#!/usr/bin/env python3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 78 |
from trl import DPOTrainer, DPOConfig
|
| 79 |
from datasets import load_dataset
|
| 80 |
|
|
|
|
| 84 |
learning_rate={lr},
|
| 85 |
num_train_epochs={epochs},
|
| 86 |
per_device_train_batch_size={batch_size},
|
| 87 |
+
max_seq_length=512,
|
| 88 |
bf16=True,
|
| 89 |
output_dir='./{output_name}-output',
|
| 90 |
push_to_hub=True,
|
|
|
|
| 100 |
trainer.train()
|
| 101 |
trainer.push_to_hub()
|
| 102 |
'''
|
| 103 |
+
else: # LoRA
|
| 104 |
+
script = f'''#!/usr/bin/env python3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 105 |
from trl import SFTTrainer, SFTConfig
|
| 106 |
from peft import LoraConfig
|
| 107 |
from datasets import load_dataset
|
|
|
|
| 116 |
)
|
| 117 |
|
| 118 |
training_args = SFTConfig(
|
| 119 |
+
learning_rate={lr * 10},
|
| 120 |
num_train_epochs={epochs},
|
| 121 |
per_device_train_batch_size={batch_size},
|
| 122 |
max_seq_length=2048,
|
|
|
|
| 136 |
trainer.train()
|
| 137 |
trainer.push_to_hub()
|
| 138 |
'''
|
| 139 |
+
|
| 140 |
+
st.code(script, language='python')
|
| 141 |
+
|
| 142 |
+
# Hardware info
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 143 |
if '0.5B' in model or '1B' in model:
|
| 144 |
+
hw = 'a10g-small (24GB VRAM) - ~2h training'
|
| 145 |
elif '3B' in model or '7B' in model:
|
| 146 |
+
hw = 'a10g-large or a100-small (24-80GB VRAM) - ~4h training'
|
| 147 |
else:
|
| 148 |
+
hw = 'a100-large (80GB VRAM) - ~6h training'
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 149 |
|
| 150 |
+
st.info(f'**Recommended:** {hw}')
|
| 151 |
+
|
| 152 |
+
st.markdown('---')
|
| 153 |
+
st.markdown('### 📋 How to Use')
|
| 154 |
+
st.markdown('''
|
| 155 |
+
1. Configure parameters above
|
| 156 |
+
2. Click **Generate Training Script**
|
| 157 |
+
3. Copy the script to `train.py`
|
| 158 |
+
4. Install: `pip install transformers trl torch datasets accelerate peft`
|
| 159 |
+
5. Run: `python train.py`
|
| 160 |
+
|
| 161 |
+
**Note:** Need Hugging Face Pro or compute credits for cloud training.
|
| 162 |
+
''')
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|