Spaces:
Paused
Paused
File size: 1,496 Bytes
d814291 fe1f842 d814291 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 | from osint_env.baselines.openai_runner import OpenAIBaselineConfig, OpenAIBaselineRunner, build_action_tools
def test_openai_baseline_toolset_contains_answer_and_graph_actions():
tools = build_action_tools()
names = {tool["function"]["name"] for tool in tools}
assert "submit_answer" in names
assert "add_edge" in names
assert "search_memory" in names
assert "search_shared_context" in names
assert "get_post" in names
def test_gpt5_request_kwargs_avoid_temperature_and_use_max_completion_tokens():
runner = OpenAIBaselineRunner.__new__(OpenAIBaselineRunner)
runner.config = OpenAIBaselineConfig(model="gpt-5-nano", max_tokens=321, temperature=0.0, seed=7)
runner.tools = build_action_tools()
kwargs = runner._request_kwargs(messages=[{"role": "user", "content": "hi"}], episode_index=0)
assert kwargs["max_completion_tokens"] == 321
assert kwargs["reasoning_effort"] == "none"
assert "temperature" not in kwargs
def test_gpt54_mini_request_kwargs_skip_reasoning_effort_for_chat_completions():
runner = OpenAIBaselineRunner.__new__(OpenAIBaselineRunner)
runner.config = OpenAIBaselineConfig(model="gpt-5.4-mini", max_tokens=321, temperature=0.0, seed=7)
runner.tools = build_action_tools()
kwargs = runner._request_kwargs(messages=[{"role": "user", "content": "hi"}], episode_index=0)
assert kwargs["max_completion_tokens"] == 321
assert "reasoning_effort" not in kwargs
assert "temperature" not in kwargs
|