BeFM / requirements.txt
Jn-Huang
Switch to vLLM for faster inference with lazy loading and multi-turn fix
89babab
raw
history blame contribute delete
76 Bytes
torch>=2.0.0
transformers>=4.30.0
peft>=0.4.0
spaces
accelerate
vllm>=0.6.0