rajkr commited on
Commit
b85a115
·
verified ·
1 Parent(s): 868ce8f

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -6
README.md CHANGED
@@ -1,12 +1,49 @@
1
  ---
2
- title: Ml Trainer Space
3
- emoji: 🌍
4
- colorFrom: gray
5
- colorTo: yellow
6
  sdk: gradio
7
- sdk_version: 6.13.0
8
  app_file: app.py
9
  pinned: false
10
  ---
11
 
12
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: ML Model Trainer
3
+ emoji: 🤖
4
+ colorFrom: blue
5
+ colorTo: green
6
  sdk: gradio
7
+ sdk_version: 4.44.0
8
  app_file: app.py
9
  pinned: false
10
  ---
11
 
12
+ # 🤖 ML Model Trainer
13
+
14
+ Free tool to generate training scripts for fine-tuning open-source LLMs.
15
+
16
+ ## Features
17
+
18
+ - **SFT** (Supervised Fine-Tuning) - Full model fine-tuning
19
+ - **DPO** (Direct Preference Optimization) - Preference alignment
20
+ - **LoRA** - Parameter-efficient fine-tuning
21
+
22
+ ## Supported Models
23
+
24
+ | Size | Models |
25
+ |------|--------|
26
+ | Small (0.5-1.5B) | Qwen2.5-0.5B, Qwen2.5-1.5B, Phi-3-mini, Gemma-2B, Llama-3.2-1B |
27
+ | Medium (7B) | Qwen2.5-7B, Llama-3.2-3B, Mistral-7B |
28
+ | Large (14B+) | Qwen2.5-14B, Mixtral-8x7B |
29
+
30
+ ## Public Datasets
31
+
32
+ - HuggingFaceH4/ultrachat_200k
33
+ - openai/gsm8k
34
+ - meta-math/MATH
35
+ - anthropic/hh-rlhf
36
+ - stanfordnlp/SHP
37
+
38
+ ## How to Use
39
+
40
+ 1. Select model, training method, and dataset
41
+ 2. Configure hyperparameters (epochs, learning rate, batch size)
42
+ 3. Generate training script
43
+ 4. Copy and run locally or on Hugging Face Jobs
44
+
45
+ ## Requirements
46
+
47
+ ```bash
48
+ pip install transformers trl torch datasets accelerate peft
49
+ ```