Spaces:
Running
Running
Update rhythma.py
Browse files- rhythma.py +1 -1
rhythma.py
CHANGED
|
@@ -429,7 +429,7 @@ State/Intention:"""
|
|
| 429 |
print(f"ℹ️ Querying Groq for emotion analysis...")
|
| 430 |
chat_completion = self.groq_client.chat.completions.create(
|
| 431 |
messages=[{"role": "user", "content": prompt}],
|
| 432 |
-
model="
|
| 433 |
max_tokens=15, # Allow slightly more tokens for flexibility
|
| 434 |
temperature=0.2, # Lower temperature for more deterministic classification
|
| 435 |
stop=["\n"] # Stop generation after the first line
|
|
|
|
| 429 |
print(f"ℹ️ Querying Groq for emotion analysis...")
|
| 430 |
chat_completion = self.groq_client.chat.completions.create(
|
| 431 |
messages=[{"role": "user", "content": prompt}],
|
| 432 |
+
model="llama-3.3-70b-versatile", # Specify a capable model available on Groq
|
| 433 |
max_tokens=15, # Allow slightly more tokens for flexibility
|
| 434 |
temperature=0.2, # Lower temperature for more deterministic classification
|
| 435 |
stop=["\n"] # Stop generation after the first line
|