rajkr commited on
Commit
8892e63
·
verified ·
1 Parent(s): a0ce63e

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -6
README.md CHANGED
@@ -3,8 +3,8 @@ title: ML Model Trainer
3
  emoji: 🤖
4
  colorFrom: blue
5
  colorTo: green
6
- sdk: gradio
7
- sdk_version: 4.44.0
8
  app_file: app.py
9
  pinned: false
10
  ---
@@ -16,16 +16,15 @@ Free tool to generate training scripts for fine-tuning open-source LLMs.
16
  ## Features
17
 
18
  - **SFT** (Supervised Fine-Tuning) - Full model fine-tuning
19
- - **DPO** (Direct Preference Optimization) - Preference alignment
20
  - **LoRA** - Parameter-efficient fine-tuning
21
 
22
  ## Supported Models
23
 
24
  | Size | Models |
25
  |------|--------|
26
- | Small (0.5-1.5B) | Qwen2.5-0.5B, Qwen2.5-1.5B, Phi-3-mini, Gemma-2B, Llama-3.2-1B |
27
  | Medium (7B) | Qwen2.5-7B, Llama-3.2-3B, Mistral-7B |
28
- | Large (14B+) | Qwen2.5-14B, Mixtral-8x7B |
29
 
30
  ## Public Datasets
31
 
@@ -33,7 +32,6 @@ Free tool to generate training scripts for fine-tuning open-source LLMs.
33
  - openai/gsm8k
34
  - meta-math/MATH
35
  - anthropic/hh-rlhf
36
- - stanfordnlp/SHP
37
 
38
  ## How to Use
39
 
 
3
  emoji: 🤖
4
  colorFrom: blue
5
  colorTo: green
6
+ sdk: streamlit
7
+ sdk_version: 1.28.0
8
  app_file: app.py
9
  pinned: false
10
  ---
 
16
  ## Features
17
 
18
  - **SFT** (Supervised Fine-Tuning) - Full model fine-tuning
19
+ - **DPO** (Direct Preference Optimization) - Preference alignment
20
  - **LoRA** - Parameter-efficient fine-tuning
21
 
22
  ## Supported Models
23
 
24
  | Size | Models |
25
  |------|--------|
26
+ | Small (0.5-1.5B) | Qwen2.5-0.5B, Qwen2.5-1.5B, Llama-3.2-1B, Phi-3-mini, Gemma-2B |
27
  | Medium (7B) | Qwen2.5-7B, Llama-3.2-3B, Mistral-7B |
 
28
 
29
  ## Public Datasets
30
 
 
32
  - openai/gsm8k
33
  - meta-math/MATH
34
  - anthropic/hh-rlhf
 
35
 
36
  ## How to Use
37