Imsachin010 commited on
Commit
f5051d6
·
1 Parent(s): 1af4cba

Fix indentation bug in grpo_train.py + update requirements.txt

Browse files
Files changed (2) hide show
  1. requirements.txt +10 -4
  2. training/grpo_train.py +2 -2
requirements.txt CHANGED
@@ -2,9 +2,15 @@ fastapi>=0.110.0
2
  uvicorn[standard]>=0.29.0
3
  pydantic>=2.0
4
  openenv-core>=0.2.3
5
- transformers
6
- datasets
7
- trl
8
- peft
9
  httpx
10
  matplotlib
 
 
 
 
 
 
 
2
  uvicorn[standard]>=0.29.0
3
  pydantic>=2.0
4
  openenv-core>=0.2.3
5
+ transformers>=4.44.0
6
+ datasets>=2.20.0
7
+ trl>=0.11.0
8
+ peft>=0.11.0
9
  httpx
10
  matplotlib
11
+ accelerate>=0.33.0
12
+ bitsandbytes>=0.43.0
13
+ huggingface_hub[cli]>=0.24.0
14
+ hf_transfer>=0.1.8
15
+ numpy<2
16
+ torch>=2.0.0
training/grpo_train.py CHANGED
@@ -38,8 +38,8 @@ def _load_model_and_tokenizer(model_name: str, use_unsloth: bool = False):
38
  if use_unsloth:
39
  try:
40
  from unsloth import FastLanguageModel
41
- print("Loading with Unsloth in 4-bit + LoRA...")
42
- model, tokenizer = FastLanguageModel.from_pretrained(
43
  model_name=model_name,
44
  max_seq_length=2048,
45
  load_in_4bit=True,
 
38
  if use_unsloth:
39
  try:
40
  from unsloth import FastLanguageModel
41
+ print("Loading with Unsloth in 4-bit + LoRA...")
42
+ model, tokenizer = FastLanguageModel.from_pretrained(
43
  model_name=model_name,
44
  max_seq_length=2048,
45
  load_in_4bit=True,