| import tempfile |
| from concurrent.futures import wait |
|
|
| import pytest |
|
|
| import gradio as gr |
| from gradio import helpers |
|
|
|
|
| def invalid_fn(message): |
| return message |
|
|
|
|
| def double(message, history): |
| return message + " " + message |
|
|
|
|
| async def async_greet(message, history): |
| return "hi, " + message |
|
|
|
|
| def stream(message, history): |
| for i in range(len(message)): |
| yield message[: i + 1] |
|
|
|
|
| async def async_stream(message, history): |
| for i in range(len(message)): |
| yield message[: i + 1] |
|
|
|
|
| def count(message, history): |
| return str(len(history)) |
|
|
|
|
| def echo_system_prompt_plus_message(message, history, system_prompt, tokens): |
| response = f"{system_prompt} {message}" |
| for i in range(min(len(response), int(tokens))): |
| yield response[: i + 1] |
|
|
|
|
| class TestInit: |
| def test_no_fn(self): |
| with pytest.raises(TypeError): |
| gr.ChatInterface() |
|
|
| def test_configuring_buttons(self): |
| chatbot = gr.ChatInterface(double, submit_btn=None, retry_btn=None) |
| assert chatbot.submit_btn is None |
| assert chatbot.retry_btn is None |
|
|
| def test_events_attached(self): |
| chatbot = gr.ChatInterface(double) |
| dependencies = chatbot.dependencies |
| textbox = chatbot.textbox._id |
| submit_btn = chatbot.submit_btn._id |
| assert next( |
| ( |
| d |
| for d in dependencies |
| if d["targets"] == [(textbox, "submit"), (submit_btn, "click")] |
| ), |
| None, |
| ) |
| for btn_id in [ |
| chatbot.retry_btn._id, |
| chatbot.clear_btn._id, |
| chatbot.undo_btn._id, |
| ]: |
| assert next( |
| (d for d in dependencies if d["targets"][0] == (btn_id, "click")), |
| None, |
| ) |
|
|
| @pytest.mark.asyncio |
| async def test_example_caching(self, monkeypatch): |
| monkeypatch.setattr(helpers, "CACHED_FOLDER", tempfile.mkdtemp()) |
| chatbot = gr.ChatInterface( |
| double, examples=["hello", "hi"], cache_examples=True |
| ) |
| prediction_hello = await chatbot.examples_handler.load_from_cache(0) |
| prediction_hi = await chatbot.examples_handler.load_from_cache(1) |
| assert prediction_hello[0][0] == ["hello", "hello hello"] |
| assert prediction_hi[0][0] == ["hi", "hi hi"] |
|
|
| @pytest.mark.asyncio |
| async def test_example_caching_async(self, monkeypatch): |
| monkeypatch.setattr(helpers, "CACHED_FOLDER", tempfile.mkdtemp()) |
| chatbot = gr.ChatInterface( |
| async_greet, examples=["abubakar", "tom"], cache_examples=True |
| ) |
| prediction_hello = await chatbot.examples_handler.load_from_cache(0) |
| prediction_hi = await chatbot.examples_handler.load_from_cache(1) |
| assert prediction_hello[0][0] == ["abubakar", "hi, abubakar"] |
| assert prediction_hi[0][0] == ["tom", "hi, tom"] |
|
|
| @pytest.mark.asyncio |
| async def test_example_caching_with_streaming(self, monkeypatch): |
| monkeypatch.setattr(helpers, "CACHED_FOLDER", tempfile.mkdtemp()) |
| chatbot = gr.ChatInterface( |
| stream, examples=["hello", "hi"], cache_examples=True |
| ) |
| prediction_hello = await chatbot.examples_handler.load_from_cache(0) |
| prediction_hi = await chatbot.examples_handler.load_from_cache(1) |
| assert prediction_hello[0][0] == ["hello", "hello"] |
| assert prediction_hi[0][0] == ["hi", "hi"] |
|
|
| @pytest.mark.asyncio |
| async def test_example_caching_with_streaming_async(self, monkeypatch): |
| monkeypatch.setattr(helpers, "CACHED_FOLDER", tempfile.mkdtemp()) |
| chatbot = gr.ChatInterface( |
| async_stream, examples=["hello", "hi"], cache_examples=True |
| ) |
| prediction_hello = await chatbot.examples_handler.load_from_cache(0) |
| prediction_hi = await chatbot.examples_handler.load_from_cache(1) |
| assert prediction_hello[0][0] == ["hello", "hello"] |
| assert prediction_hi[0][0] == ["hi", "hi"] |
|
|
| @pytest.mark.asyncio |
| async def test_example_caching_with_additional_inputs(self, monkeypatch): |
| monkeypatch.setattr(helpers, "CACHED_FOLDER", tempfile.mkdtemp()) |
| chatbot = gr.ChatInterface( |
| echo_system_prompt_plus_message, |
| additional_inputs=["textbox", "slider"], |
| examples=[["hello", "robot", 100], ["hi", "robot", 2]], |
| cache_examples=True, |
| ) |
| prediction_hello = await chatbot.examples_handler.load_from_cache(0) |
| prediction_hi = await chatbot.examples_handler.load_from_cache(1) |
| assert prediction_hello[0][0] == ["hello", "robot hello"] |
| assert prediction_hi[0][0] == ["hi", "ro"] |
|
|
| @pytest.mark.asyncio |
| async def test_example_caching_with_additional_inputs_already_rendered( |
| self, monkeypatch |
| ): |
| monkeypatch.setattr(helpers, "CACHED_FOLDER", tempfile.mkdtemp()) |
| with gr.Blocks(): |
| with gr.Accordion("Inputs"): |
| text = gr.Textbox() |
| slider = gr.Slider() |
| chatbot = gr.ChatInterface( |
| echo_system_prompt_plus_message, |
| additional_inputs=[text, slider], |
| examples=[["hello", "robot", 100], ["hi", "robot", 2]], |
| cache_examples=True, |
| ) |
| prediction_hello = await chatbot.examples_handler.load_from_cache(0) |
| prediction_hi = await chatbot.examples_handler.load_from_cache(1) |
| assert prediction_hello[0][0] == ["hello", "robot hello"] |
| assert prediction_hi[0][0] == ["hi", "ro"] |
|
|
|
|
| class TestAPI: |
| def test_get_api_info(self): |
| chatbot = gr.ChatInterface(double) |
| api_info = gr.blocks.get_api_info(chatbot.get_config_file()) |
| assert len(api_info["named_endpoints"]) == 1 |
| assert len(api_info["unnamed_endpoints"]) == 0 |
| assert "/chat" in api_info["named_endpoints"] |
|
|
| def test_streaming_api(self, connect): |
| chatbot = gr.ChatInterface(stream).queue() |
| with connect(chatbot) as client: |
| job = client.submit("hello") |
| wait([job]) |
| assert job.outputs() == ["h", "he", "hel", "hell", "hello"] |
|
|
| def test_streaming_api_async(self, connect): |
| chatbot = gr.ChatInterface(async_stream).queue() |
| with connect(chatbot) as client: |
| job = client.submit("hello") |
| wait([job]) |
| assert job.outputs() == ["h", "he", "hel", "hell", "hello"] |
|
|
| def test_non_streaming_api(self, connect): |
| chatbot = gr.ChatInterface(double) |
| with connect(chatbot) as client: |
| result = client.predict("hello") |
| assert result == "hello hello" |
|
|
| def test_non_streaming_api_async(self, connect): |
| chatbot = gr.ChatInterface(async_greet) |
| with connect(chatbot) as client: |
| result = client.predict("gradio") |
| assert result == "hi, gradio" |
|
|
| def test_streaming_api_with_additional_inputs(self, connect): |
| chatbot = gr.ChatInterface( |
| echo_system_prompt_plus_message, |
| additional_inputs=["textbox", "slider"], |
| ).queue() |
| with connect(chatbot) as client: |
| job = client.submit("hello", "robot", 7) |
| wait([job]) |
| assert job.outputs() == [ |
| "r", |
| "ro", |
| "rob", |
| "robo", |
| "robot", |
| "robot ", |
| "robot h", |
| ] |
|
|