File size: 1,496 Bytes
d814291
 
 
 
 
 
 
 
 
fe1f842
d814291
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
from osint_env.baselines.openai_runner import OpenAIBaselineConfig, OpenAIBaselineRunner, build_action_tools


def test_openai_baseline_toolset_contains_answer_and_graph_actions():
    tools = build_action_tools()
    names = {tool["function"]["name"] for tool in tools}
    assert "submit_answer" in names
    assert "add_edge" in names
    assert "search_memory" in names
    assert "search_shared_context" in names
    assert "get_post" in names


def test_gpt5_request_kwargs_avoid_temperature_and_use_max_completion_tokens():
    runner = OpenAIBaselineRunner.__new__(OpenAIBaselineRunner)
    runner.config = OpenAIBaselineConfig(model="gpt-5-nano", max_tokens=321, temperature=0.0, seed=7)
    runner.tools = build_action_tools()
    kwargs = runner._request_kwargs(messages=[{"role": "user", "content": "hi"}], episode_index=0)
    assert kwargs["max_completion_tokens"] == 321
    assert kwargs["reasoning_effort"] == "none"
    assert "temperature" not in kwargs


def test_gpt54_mini_request_kwargs_skip_reasoning_effort_for_chat_completions():
    runner = OpenAIBaselineRunner.__new__(OpenAIBaselineRunner)
    runner.config = OpenAIBaselineConfig(model="gpt-5.4-mini", max_tokens=321, temperature=0.0, seed=7)
    runner.tools = build_action_tools()
    kwargs = runner._request_kwargs(messages=[{"role": "user", "content": "hi"}], episode_index=0)
    assert kwargs["max_completion_tokens"] == 321
    assert "reasoning_effort" not in kwargs
    assert "temperature" not in kwargs