| |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| from google.cloud.aiplatform import base |
| from google.cloud.aiplatform import initializer as aiplatform_initializer |
| from google.cloud.aiplatform.compat.types import dataset as gca_dataset |
| from google.cloud.aiplatform_v1.types import ( |
| dataset_version as gca_dataset_version, |
| ) |
| from google.cloud.aiplatform_v1beta1.types import ( |
| prediction_service as gapic_prediction_service_types, |
| ) |
| from vertexai.generative_models import ( |
| Part, |
| Image, |
| GenerativeModel, |
| Tool, |
| ToolConfig, |
| ) |
| from vertexai.generative_models._generative_models import ( |
| _proto_to_dict, |
| _dict_to_proto, |
| _tool_types_to_gapic_tools, |
| PartsType, |
| ) |
| from vertexai.prompts._prompts import Prompt |
| from google.protobuf import field_mask_pb2 as field_mask |
|
|
| import dataclasses |
| from typing import ( |
| Any, |
| Dict, |
| Optional, |
| ) |
|
|
| _LOGGER = base.Logger(__name__) |
| _dataset_client_value = None |
|
|
| DEFAULT_API_SCHEMA_VERSION = "1.0.0" |
| PROMPT_SCHEMA_URI = ( |
| "gs://google-cloud-aiplatform/schema/dataset/metadata/text_prompt_1.0.0.yaml" |
| ) |
|
|
|
|
| def _format_function_declaration_parameters(obj: Any): |
| """Recursively replaces type_ and format_ fields in-place.""" |
| if isinstance(obj, (str, int, float)): |
| return obj |
| if isinstance(obj, dict): |
| new = obj.__class__() |
| for key, value in obj.items(): |
| key = key.replace("type_", "type") |
| key = key.replace("format_", "format") |
| new[key] = _format_function_declaration_parameters(value) |
| elif isinstance(obj, (list, set, tuple)): |
| new = obj.__class__( |
| _format_function_declaration_parameters(value) for value in obj |
| ) |
| else: |
| return obj |
| return new |
|
|
|
|
| @dataclasses.dataclass |
| class Arguments: |
| """Arguments. Child of Execution. |
| |
| Attributes: |
| variables: The arguments of the execution. |
| """ |
|
|
| variables: dict[str, list[Part]] |
|
|
| def to_dict(self) -> Dict[str, Any]: |
| dct = {} |
| for variable_name in self.variables: |
| dct[variable_name] = { |
| "partList": { |
| "parts": [part.to_dict() for part in self.variables[variable_name]] |
| } |
| } |
| return dct |
|
|
| @classmethod |
| def from_dict(cls, dct: Dict[str, Any]) -> "Arguments": |
| variables = {} |
| for variable_name in dct: |
| variables[variable_name] = [ |
| Part.from_dict(part) for part in dct[variable_name]["partList"]["parts"] |
| ] |
| arguments = cls(variables=variables) |
| return arguments |
|
|
|
|
| @dataclasses.dataclass |
| class Execution: |
| """Execution. Child of MultimodalPrompt. |
| |
| Attributes: |
| arguments: The arguments of the execution. |
| """ |
|
|
| arguments: Arguments |
|
|
| def __init__(self, arguments: dict[str, list[Part]]): |
| self.arguments = Arguments(variables=arguments) |
|
|
| def to_dict(self) -> Dict[str, Any]: |
| dct = {} |
| dct["arguments"] = self.arguments.to_dict() |
| return dct |
|
|
| @classmethod |
| def from_dict(cls, dct: Dict[str, Any]) -> "Execution": |
| arguments = dct.get("arguments", None) |
| execution = cls(arguments=arguments) |
| return execution |
|
|
|
|
| @dataclasses.dataclass |
| class MultimodalPrompt: |
| """MultimodalPrompt. Child of PromptDatasetMetadata. |
| |
| Attributes: |
| prompt_message: The schema for the prompt. A subset of the GenerateContentRequest schema. |
| api_schema_version: The api schema version of the prompt when it was last modified. |
| executions: Contains data related to an execution of a prompt (ex. variables) |
| """ |
|
|
| prompt_message: gapic_prediction_service_types.GenerateContentRequest |
| api_schema_version: Optional[str] = DEFAULT_API_SCHEMA_VERSION |
| executions: Optional[list[Execution]] = None |
|
|
| def to_dict(self) -> Dict[str, Any]: |
| dct = {"multimodalPrompt": {}} |
| dct["apiSchemaVersion"] = self.api_schema_version |
| dct["multimodalPrompt"]["promptMessage"] = _proto_to_dict(self.prompt_message) |
|
|
| |
| if dct["multimodalPrompt"]["promptMessage"].get("tools", None): |
| tools = dct["multimodalPrompt"]["promptMessage"]["tools"] |
| for tool in tools: |
| for function_declaration in tool.get("function_declarations", []): |
| function_declaration[ |
| "parameters" |
| ] = _format_function_declaration_parameters( |
| function_declaration["parameters"] |
| ) |
|
|
| if self.executions and self.executions[0]: |
| |
| execution_dcts = [] |
| for execution in self.executions: |
| exeuction_dct = execution.to_dict() |
| if exeuction_dct and exeuction_dct["arguments"]: |
| execution_dcts.append(exeuction_dct) |
| if execution_dcts: |
| dct["executions"] = execution_dcts |
| return dct |
|
|
| @classmethod |
| def from_dict(cls, dct: Dict[str, Any]) -> "MultimodalPrompt": |
| api_schema_version = dct.get("apiSchemaVersion", DEFAULT_API_SCHEMA_VERSION) |
| if int(api_schema_version.split(".")[0]) > int( |
| DEFAULT_API_SCHEMA_VERSION.split(".")[0] |
| ): |
| |
| raise ValueError( |
| "This prompt was saved with a newer schema version and cannot be loaded." |
| ) |
| prompt_message_dct = dct.get("multimodalPrompt", {}).get("promptMessage", None) |
| if not prompt_message_dct: |
| raise ValueError("This prompt is not supported in the SDK.") |
| |
| tools = prompt_message_dct.get("tools", None) |
| if tools: |
| tools = [Tool.from_dict(tool) for tool in tools] |
| prompt_message_dct.pop("tools") |
| prompt_message = _dict_to_proto( |
| gapic_prediction_service_types.GenerateContentRequest, prompt_message_dct |
| ) |
| if tools: |
| |
| prompt_message.tools = _tool_types_to_gapic_tools(tools) |
| executions_dct = dct.get("executions", []) |
| executions = [Execution.from_dict(execution) for execution in executions_dct] |
| if not executions: |
| executions = None |
| multimodal_prompt = cls( |
| prompt_message=prompt_message, |
| api_schema_version=api_schema_version, |
| executions=executions, |
| ) |
| return multimodal_prompt |
|
|
|
|
| @dataclasses.dataclass |
| class PromptDatasetMetadata: |
| """PromptDatasetMetadata. |
| |
| Attributes: |
| prompt_type: Required. SDK only supports "freeform" or "multimodal_freeform" |
| prompt_api_schema: Required. SDK only supports multimodalPrompt |
| """ |
|
|
| prompt_type: str |
| prompt_api_schema: MultimodalPrompt |
|
|
| def to_dict(self) -> Dict[str, Any]: |
| dct = {} |
| dct["promptType"] = self.prompt_type |
| dct["promptApiSchema"] = self.prompt_api_schema.to_dict() |
| return dct |
|
|
| @classmethod |
| def from_dict(cls, dct: Dict[str, Any]) -> "PromptDatasetMetadata": |
| metadata = cls( |
| prompt_type=dct.get("promptType", None), |
| prompt_api_schema=MultimodalPrompt.from_dict( |
| dct.get("promptApiSchema", None) |
| ), |
| ) |
| return metadata |
|
|
|
|
| @dataclasses.dataclass |
| class PromptMetadata: |
| """Metadata containing the display name and prompt id of a prompt. |
| |
| Returned by the `list_prompts` method. |
| |
| Attributes: |
| display_name: The display name of the prompt version. |
| prompt_id: The id of the prompt. |
| """ |
|
|
| display_name: str |
| prompt_id: str |
|
|
|
|
| @dataclasses.dataclass |
| class PromptVersionMetadata: |
| """Metadata containing the display name, prompt id, and version id of a prompt version. |
| |
| Returned by the `list_prompt_versions` method. |
| |
| Attributes: |
| display_name: The display name of the prompt version. |
| prompt_id: The id of the prompt. |
| version_id: The version id of the prompt. |
| """ |
|
|
| display_name: str |
| prompt_id: str |
| version_id: str |
|
|
|
|
| def create_version( |
| prompt: Prompt, |
| prompt_id: Optional[str] = None, |
| version_name: Optional[str] = None, |
| ) -> Prompt: |
| """Creates a Prompt or Prompt Version in the online prompt store |
| |
| Args: |
| prompt: The Prompt object to create a new version of. |
| prompt_id: The id of the prompt resource to create a new version under. |
| If it is not provided and the prompt has no prompt resource |
| associated with it, a new prompt resource will be created. |
| version_name: Optional display name of the new prompt version. |
| If not specified, a default name including a timestamp will be used. |
| |
| Returns: |
| A new Prompt object with a reference to the newly created or updated |
| prompt resource. This new Prompt object is nearly identical to the |
| original Prompt object, except it has references to the new |
| prompt version. |
| """ |
| if not (prompt_id or prompt._dataset): |
| |
| |
| return _create_prompt_resource(prompt=prompt, version_name=version_name) |
|
|
| |
| if not prompt_id: |
| return _create_prompt_version_resource(prompt=prompt, version_name=version_name) |
|
|
| |
| |
|
|
| |
| original_prompt_resource = None if not prompt._dataset else prompt._dataset.name |
|
|
| |
| if not prompt._dataset: |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| name = f"projects/{project}/locations/{location}/datasets/{prompt_id}" |
| dataset_metadata = _format_dataset_metadata_dict(prompt=prompt) |
|
|
| prompt._dataset = gca_dataset.Dataset( |
| name=name, |
| display_name=prompt.prompt_name or "Untitled Prompt", |
| metadata_schema_uri=PROMPT_SCHEMA_URI, |
| metadata=dataset_metadata, |
| model_reference=prompt.model_name, |
| ) |
|
|
| |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| prompt._dataset.name = ( |
| f"projects/{project}/locations/{location}/datasets/{prompt_id}" |
| ) |
| result = _create_prompt_version_resource(prompt=prompt, version_name=version_name) |
|
|
| |
| |
| prompt._dataset.name = original_prompt_resource |
| return result |
|
|
|
|
| def _check_multimodal_contents(prompt_data: PartsType): |
| if isinstance(prompt_data, Image): |
| return "multimodal_freeform" |
| elif isinstance(prompt_data, list): |
| for part in prompt_data: |
| check = _check_multimodal_contents(part) |
| if check == "multimodal_freeform": |
| return "multimodal_freeform" |
| elif isinstance(prompt_data, Part): |
| if "text" not in prompt_data._raw_part: |
| return "multimodal_freeform" |
| return "freeform" |
|
|
|
|
| def _format_dataset_metadata_dict(prompt: Prompt) -> dict[str, Any]: |
| """Helper function to convert the configs and prompt data stored in the Prompt object to a dataset metadata dict.""" |
| model = GenerativeModel(model_name=prompt.model_name) |
| prompt_message = model._prepare_request( |
| contents=prompt.prompt_data or "temporary data", |
| model=prompt.model_name, |
| system_instruction=prompt.system_instruction, |
| tools=prompt.tools, |
| tool_config=prompt.tool_config, |
| safety_settings=prompt.safety_settings, |
| generation_config=prompt.generation_config, |
| ) |
| |
| if not prompt.prompt_data: |
| prompt_message.contents = None |
|
|
| |
| if prompt.prompt_data: |
| prompt_type = _check_multimodal_contents(prompt.prompt_data) |
| else: |
| prompt_type = "freeform" |
|
|
| return PromptDatasetMetadata( |
| prompt_type=prompt_type, |
| prompt_api_schema=MultimodalPrompt( |
| prompt_message=prompt_message, |
| executions=[Execution(variable_set) for variable_set in prompt.variables], |
| ), |
| ).to_dict() |
|
|
|
|
| def _create_dataset(prompt: Prompt, parent: str) -> gca_dataset.Dataset: |
| dataset_metadata = _format_dataset_metadata_dict(prompt=prompt) |
| dataset = gca_dataset.Dataset( |
| name=parent, |
| display_name=prompt.prompt_name or "Untitled Prompt", |
| metadata_schema_uri=PROMPT_SCHEMA_URI, |
| metadata=dataset_metadata, |
| model_reference=prompt.model_name, |
| ) |
| operation = prompt._dataset_client.create_dataset( |
| parent=parent, |
| dataset=dataset, |
| ) |
| dataset = operation.result() |
|
|
| |
| dataset.labels = None |
| return dataset |
|
|
|
|
| def _create_dataset_version( |
| prompt: Prompt, parent: str, version_name: Optional[str] = None |
| ): |
| dataset_version = gca_dataset_version.DatasetVersion( |
| display_name=version_name, |
| ) |
|
|
| dataset_version = prompt._dataset_client.create_dataset_version( |
| parent=parent, |
| dataset_version=dataset_version, |
| ) |
| return dataset_version.result() |
|
|
|
|
| def _update_dataset( |
| prompt: Prompt, |
| dataset: gca_dataset.Dataset, |
| ) -> gca_dataset_version.DatasetVersion: |
| dataset.metadata = _format_dataset_metadata_dict(prompt=prompt) |
|
|
| mask_paths = ["modelReference", "metadata"] |
| if dataset.display_name != "Untitled Prompt": |
| mask_paths.append("displayName") |
|
|
| updated_dataset = prompt._dataset_client.update_dataset( |
| dataset=dataset, |
| update_mask=field_mask.FieldMask(paths=mask_paths), |
| ) |
| |
| updated_dataset.etag = None |
| return updated_dataset |
|
|
|
|
| def _create_prompt_resource( |
| prompt: Prompt, version_name: Optional[str] = None |
| ) -> Prompt: |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
|
|
| |
| parent = f"projects/{project}/locations/{location}" |
| dataset = _create_dataset(prompt=prompt, parent=parent) |
|
|
| |
| dataset_version = _create_dataset_version( |
| prompt=prompt, |
| parent=dataset.name, |
| version_name=version_name, |
| ) |
|
|
| |
| new_prompt = Prompt._clone(prompt=prompt) |
| new_prompt._dataset = dataset |
| new_prompt._version_id = dataset_version.name.split("/")[-1] |
| new_prompt._version_name = dataset_version.display_name |
| prompt_id = new_prompt._dataset.name.split("/")[5] |
|
|
| _LOGGER.info( |
| f"Created prompt resource with id {prompt_id} with version number {new_prompt._version_id}" |
| ) |
| return new_prompt |
|
|
|
|
| def _create_prompt_version_resource( |
| prompt: Prompt, |
| version_name: Optional[str] = None, |
| ) -> Prompt: |
| |
| updated_dataset = _update_dataset(prompt=prompt, dataset=prompt._dataset) |
|
|
| |
| dataset_version = _create_dataset_version( |
| prompt=prompt, |
| parent=updated_dataset.name, |
| version_name=version_name, |
| ) |
|
|
| |
| new_prompt = Prompt._clone(prompt=prompt) |
| new_prompt._dataset = updated_dataset |
| new_prompt._version_id = dataset_version.name.split("/")[-1] |
| new_prompt._version_name = dataset_version.display_name |
| prompt_id = prompt._dataset.name.split("/")[5] |
|
|
| _LOGGER.info( |
| f"Updated prompt resource with id {prompt_id} as version number {new_prompt._version_id}" |
| ) |
| return new_prompt |
|
|
|
|
| def _get_prompt_resource(prompt: Prompt, prompt_id: str) -> gca_dataset.Dataset: |
| """Helper function to get a prompt resource from a prompt id.""" |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| name = f"projects/{project}/locations/{location}/datasets/{prompt_id}" |
| dataset = prompt._dataset_client.get_dataset(name=name) |
| return dataset |
|
|
|
|
| def _get_prompt_resource_from_version( |
| prompt: Prompt, prompt_id: str, version_id: str |
| ) -> gca_dataset.Dataset: |
| """Helper function to get a prompt resource from a prompt version id.""" |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| name = f"projects/{project}/locations/{location}/datasets/{prompt_id}/datasetVersions/{version_id}" |
|
|
| |
| dataset_version = prompt._dataset_client.get_dataset_version(name=name) |
| prompt._version_name = dataset_version.display_name |
|
|
| |
| name = f"projects/{project}/locations/{location}/datasets/{prompt_id}" |
| dataset = prompt._dataset_client.get_dataset(name=name) |
|
|
| |
| dataset = gca_dataset.Dataset( |
| name=name, |
| display_name=dataset.display_name, |
| metadata_schema_uri=PROMPT_SCHEMA_URI, |
| metadata=dataset_version.metadata, |
| model_reference=dataset_version.model_reference, |
| ) |
| return dataset |
|
|
|
|
| def restore_version(prompt_id: str, version_id: str) -> PromptVersionMetadata: |
| """Restores a previous version of the prompt resource and |
| loads that version into the current Prompt object. |
| |
| Args: |
| prompt_id: The id of the prompt resource to restore a version of. |
| version_id: The version id of the online prompt resource. |
| """ |
|
|
| |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| name = f"projects/{project}/locations/{location}/datasets/{prompt_id}/datasetVersions/{version_id}" |
|
|
| |
| temp_prompt = Prompt() |
| operation = temp_prompt._dataset_client.restore_dataset_version(name=name) |
| result = operation.result() |
| new_version_id = result.name.split("/")[-1] |
| prompt_id = result.name.split("/")[5] |
|
|
| _LOGGER.info( |
| f"Restored prompt version {version_id} under prompt id {prompt_id} as version number {new_version_id}" |
| ) |
|
|
| |
| return PromptVersionMetadata( |
| display_name=result.display_name, |
| prompt_id=result.name.split("/")[5], |
| version_id=new_version_id, |
| ) |
|
|
|
|
| def get(prompt_id: str, version_id: Optional[str] = None) -> Prompt: |
| """Creates a Prompt object from an online resource. |
| |
| Args: |
| prompt_id: The id of the prompt resource. |
| version_id: Optional version id of the prompt resource. |
| If not specified, the latest version will be used. |
| |
| Returns: |
| A prompt loaded from the online resource as a `Prompt` object. |
| """ |
| prompt = Prompt() |
| if version_id: |
| dataset = _get_prompt_resource_from_version( |
| prompt=prompt, |
| prompt_id=prompt_id, |
| version_id=version_id, |
| ) |
| else: |
| dataset = _get_prompt_resource(prompt=prompt, prompt_id=prompt_id) |
|
|
| |
| dataset.etag = None |
|
|
| prompt._dataset = dataset |
| prompt._version_id = version_id |
|
|
| dataset_dict = _proto_to_dict(dataset) |
|
|
| metadata = PromptDatasetMetadata.from_dict(dataset_dict["metadata"]) |
| _populate_fields_from_metadata(prompt=prompt, metadata=metadata) |
| return prompt |
|
|
|
|
| def _populate_fields_from_metadata( |
| prompt: Prompt, metadata: PromptDatasetMetadata |
| ) -> None: |
| """Helper function to populate Promptfields from metadata object""" |
| |
| prompt.model_name = metadata.prompt_api_schema.prompt_message.model |
|
|
| |
| contents = metadata.prompt_api_schema.prompt_message.contents |
| if contents: |
| if len(contents) > 1: |
| raise ValueError("Multi-turn prompts are not supported yet.") |
| prompt_data = [Part._from_gapic(part) for part in list(contents[0].parts)] |
|
|
| |
| if len(prompt_data) == 1 and "text" in prompt_data[0]._raw_part: |
| prompt.prompt_data = prompt_data[0].text |
| else: |
| prompt.prompt_data = prompt_data |
|
|
| |
| system_instruction = metadata.prompt_api_schema.prompt_message.system_instruction |
| if system_instruction: |
| system_instruction_parts = [ |
| Part._from_gapic(part) for part in list(system_instruction.parts) |
| ] |
| |
| if len(system_instruction_parts) == 1 and system_instruction_parts[0].text: |
| prompt.system_instruction = system_instruction_parts[0].text |
| else: |
| prompt.system_instruction = system_instruction_parts |
|
|
| |
| executions = metadata.prompt_api_schema.executions |
| variables = [] |
| if executions: |
| for execution in executions: |
| serialized_variable_set = execution.arguments |
| variable_set = {} |
| if serialized_variable_set: |
| for name, value in serialized_variable_set.variables.items(): |
| |
| variable_set[name] = [ |
| Part.from_dict(part) |
| for part in list(value["partList"]["parts"]) |
| ] |
| variables.append(variable_set) |
|
|
| |
| for variable_set in variables: |
| for name, value in variable_set.items(): |
| if len(value) == 1 and "text" in value[0]._raw_part: |
| variable_set[name] = value[0].text |
| prompt.variables = variables |
|
|
| |
| generation_config = metadata.prompt_api_schema.prompt_message.generation_config |
| if generation_config: |
| prompt.generation_config = generation_config |
|
|
| |
| safety_settings = metadata.prompt_api_schema.prompt_message.safety_settings |
| if safety_settings: |
| prompt.safety_settings = list(safety_settings) |
|
|
| |
| tools = metadata.prompt_api_schema.prompt_message.tools |
| if tools: |
| prompt.tools = list(tools) |
|
|
| |
| tool_config = metadata.prompt_api_schema.prompt_message.tool_config |
| if tool_config: |
| prompt.tool_config = ToolConfig._from_gapic(tool_config) |
|
|
|
|
| def list_prompts() -> list[PromptMetadata]: |
| """Lists all prompt resources in the online prompt store associated with the project.""" |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| parent = f"projects/{project}/locations/{location}" |
|
|
| |
| temp_prompt = Prompt() |
| prompts_pager = temp_prompt._dataset_client.list_datasets( |
| parent=parent, |
| ) |
| prompts_list = [] |
| for prompt in prompts_pager: |
| prompts_list.append( |
| PromptMetadata( |
| display_name=prompt.display_name, |
| prompt_id=prompt.name.split("/")[5], |
| ) |
| ) |
| return prompts_list |
|
|
|
|
| def list_versions(prompt_id: str) -> list[PromptVersionMetadata]: |
| """Returns a list of PromptVersionMetadata objects for the prompt resource. |
| |
| Args: |
| prompt_id: The id of the prompt resource to list versions of. |
| |
| Returns: |
| A list of PromptVersionMetadata objects for the prompt resource. |
| """ |
| |
| temp_prompt = Prompt() |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| parent = f"projects/{project}/locations/{location}/datasets/{prompt_id}" |
|
|
| versions_pager = temp_prompt._dataset_client.list_dataset_versions( |
| parent=parent, |
| ) |
| version_history = [] |
| for version in versions_pager: |
| version_history.append( |
| PromptVersionMetadata( |
| display_name=version.display_name, |
| prompt_id=version.name.split("/")[5], |
| version_id=version.name.split("/")[-1], |
| ) |
| ) |
| return version_history |
|
|
|
|
| def delete(prompt_id: str) -> None: |
| """Deletes the online prompt resource associated with the prompt id.""" |
|
|
| |
| temp_prompt = Prompt() |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| name = f"projects/{project}/locations/{location}/datasets/{prompt_id}" |
|
|
| operation = temp_prompt._dataset_client.delete_dataset( |
| name=name, |
| ) |
| operation.result() |
|
|
| _LOGGER.info(f"Deleted prompt resource with id {prompt_id}.") |
|
|