hc99's picture
Add files using upload-large-folder tool
1856027 verified
raw
history blame
6.86 kB
# -*- coding: utf-8 -*-
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
# pylint: disable=protected-access, g-multiple-import
"""System tests for GenAI prompts."""
from google.cloud import aiplatform
from vertexai import generative_models
from vertexai.generative_models import (
GenerationConfig,
SafetySetting,
ToolConfig,
)
from vertexai.preview import prompts
from vertexai.preview.prompts import Prompt
from tests.system.aiplatform import e2e_base
from google import auth
_REQUEST_FUNCTION_PARAMETER_SCHEMA_STRUCT = {
"type": "object",
"properties": {
"location": {
"type": "string",
"description": "The city and state, e.g. San Francisco, CA",
},
"unit": {
"type": "string",
"enum": [
"celsius",
"fahrenheit",
],
},
},
"required": ["location"],
}
class TestPrompts(e2e_base.TestEndToEnd):
"""System tests for prompts."""
_temp_prefix = "temp_prompts_test_"
def setup_method(self):
super().setup_method()
credentials, _ = auth.default(
scopes=["https://www.googleapis.com/auth/cloud-platform"]
)
aiplatform.init(
project=e2e_base._PROJECT,
location=e2e_base._LOCATION,
credentials=credentials,
)
def test_create_prompt_with_variables(self):
# Create local Prompt
prompt = Prompt(
prompt_data="Hello, {name}! Today is {day}. How are you?",
variables=[
{"name": "Alice", "day": "Monday"},
{"name": "Bob", "day": "Tuesday"},
],
generation_config=GenerationConfig(temperature=0.1),
model_name="gemini-1.0-pro-002",
safety_settings=[
SafetySetting(
category=SafetySetting.HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT,
threshold=SafetySetting.HarmBlockThreshold.BLOCK_MEDIUM_AND_ABOVE,
method=SafetySetting.HarmBlockMethod.SEVERITY,
)
],
system_instruction="Please answer in a short sentence.",
)
# Generate content using the assembled prompt for each variable set.
for i in range(len(prompt.variables)):
prompt.generate_content(
contents=prompt.assemble_contents(**prompt.variables[i])
)
# Save Prompt to online resource. Returns a new Prompt object associated with the online resource
prompt1 = prompts.create_version(prompt=prompt)
# Only new prompt should be associated with a prompt resource
assert prompt1.prompt_id
assert not prompt.prompt_id
# Update prompt and save a new version
prompt1.prompt_data = "Hi, {name}! How are you? Today is {day}."
prompt2 = prompts.create_version(prompt=prompt1, version_name="v2")
assert prompt2.prompt_id == prompt1.prompt_id
assert prompt2.version_id != prompt1.version_id
# Restore previous version
metadata = prompts.restore_version(
prompt_id=prompt2.prompt_id, version_id=prompt1.version_id
)
assert metadata.prompt_id == prompt2.prompt_id
assert metadata.version_id != prompt2.version_id
# List prompt versions
versions_metadata = prompts.list_versions(prompt_id=metadata.prompt_id)
assert len(versions_metadata) == 3
# Delete the prompt resource
prompts.delete(prompt_id=prompt2.prompt_id)
def test_create_prompt_with_function_calling(self):
# Create local Prompt
get_current_weather_func = generative_models.FunctionDeclaration(
name="get_current_weather",
description="Get the current weather in a given location",
parameters=_REQUEST_FUNCTION_PARAMETER_SCHEMA_STRUCT,
)
weather_tool = generative_models.Tool(
function_declarations=[get_current_weather_func],
)
tool_config = ToolConfig(
function_calling_config=ToolConfig.FunctionCallingConfig(
mode=ToolConfig.FunctionCallingConfig.Mode.ANY,
allowed_function_names=["get_current_weather"],
)
)
prompt = Prompt(
prompt_data="What is the weather like in Boston?",
tools=[weather_tool],
tool_config=tool_config,
model_name="gemini-1.0-pro-002",
)
# (Optional) Create a separate prompt resource to save the version to
prompt_temp = Prompt(model_name="gemini-1.0-pro-002")
prompt_temp1 = prompts.create_version(prompt=prompt_temp, version_name="empty")
# Create a new version to an existing prompt
prompt1 = prompts.create_version(
prompt=prompt, prompt_id=prompt_temp1.prompt_id, version_name="fc"
)
# Delete the prompt resource
prompts.delete(prompt_id=prompt1.prompt_id)
def test_get_prompt_with_variables(self):
# List prompts
prompts_list = prompts.list()
assert prompts_list
# Get prompt created in UI
prompt_id = "3217694940163211264"
prompt = prompts.get(prompt_id=prompt_id)
assert prompt.prompt_id == prompt_id
assert prompt.prompt_data
assert prompt.generation_config
assert prompt.system_instruction
# UI has a bug where safety settings are not saved
# assert prompt.safety_settings
# Generate content using the assembled prompt for each variable set.
for i in range(len(prompt.variables)):
response = prompt.generate_content(
contents=prompt.assemble_contents(**prompt.variables[i])
)
assert response.text
def test_get_prompt_with_function_calling(self):
# List prompts
prompts_list = prompts.list()
assert prompts_list
# Get prompt created in UI
prompt_id = "1173060709337006080"
prompt = prompts.get(prompt_id=prompt_id)
assert prompt.prompt_id == prompt_id
assert prompt.tools
# Generate content using the prompt
response = prompt.generate_content(contents=prompt.assemble_contents())
assert response