Spaces:
Sleeping
Sleeping
| from langchain_community.llms import HuggingFaceHub | |
| def llm_node(question, search_result): | |
| # Initialize Hugging Face model (free) | |
| llm = HuggingFaceHub( | |
| repo_id="HuggingFaceH4/zephyr-7b-beta", # You can swap with mistral or mixtral if needed | |
| model_kwargs={ | |
| "temperature": 0.1, | |
| "max_new_tokens": 500 | |
| } | |
| ) | |
| # Build prompt combining search + question | |
| prompt = f"""You are solving a GAIA benchmark evaluation question. | |
| Here’s the question: | |
| {question} | |
| Here’s retrieved information: | |
| {search_result} | |
| ⚠️ VERY IMPORTANT: | |
| - ONLY return the final answer, exactly as required. | |
| - Do NOT include explanations, prefixes, or notes. | |
| - If the question asks for a list, give only the list, in the requested format. | |
| Your answer:""" | |
| response = llm.invoke(prompt) | |
| return response.strip() | |