Update app.py
Browse files
app.py
CHANGED
|
@@ -9,7 +9,7 @@ model = Llama.from_pretrained(
|
|
| 9 |
repo_id="unsloth/Qwen3.5-35B-A3B-GGUF",
|
| 10 |
filename="Qwen3.5-35B-A3B-UD-IQ4_XS.gguf",
|
| 11 |
n_ctx=2048,
|
| 12 |
-
n_threads=
|
| 13 |
)
|
| 14 |
|
| 15 |
def get_stats():
|
|
@@ -22,9 +22,6 @@ def get_stats():
|
|
| 22 |
|
| 23 |
def chat(message, history):
|
| 24 |
messages = [{"role": "system", "content": "Reply directly without any reasoning or thinking process."}]
|
| 25 |
-
for user, assistant in history:
|
| 26 |
-
messages.append({"role": "user", "content": user})
|
| 27 |
-
messages.append({"role": "assistant", "content": assistant})
|
| 28 |
messages.append({"role": "user", "content": message})
|
| 29 |
|
| 30 |
output = ""
|
|
|
|
| 9 |
repo_id="unsloth/Qwen3.5-35B-A3B-GGUF",
|
| 10 |
filename="Qwen3.5-35B-A3B-UD-IQ4_XS.gguf",
|
| 11 |
n_ctx=2048,
|
| 12 |
+
n_threads=8,
|
| 13 |
)
|
| 14 |
|
| 15 |
def get_stats():
|
|
|
|
| 22 |
|
| 23 |
def chat(message, history):
|
| 24 |
messages = [{"role": "system", "content": "Reply directly without any reasoning or thinking process."}]
|
|
|
|
|
|
|
|
|
|
| 25 |
messages.append({"role": "user", "content": message})
|
| 26 |
|
| 27 |
output = ""
|