| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import dataclasses |
| from typing import ( |
| Iterable, |
| List, |
| Sequence, |
| Optional, |
| Union, |
| ) |
|
|
| from vertexai.generative_models._generative_models import ( |
| ContentsType, |
| Image, |
| Tool, |
| PartsType, |
| _validate_contents_type_as_valid_sequence, |
| _content_types_to_gapic_contents, |
| _to_content, |
| ) |
|
|
| from vertexai.tokenization._tokenizer_loading import ( |
| get_sentencepiece, |
| get_tokenizer_name, |
| load_model_proto, |
| ) |
| from google.cloud.aiplatform_v1beta1.types import ( |
| content as gapic_content_types, |
| tool as gapic_tool_types, |
| openapi, |
| ) |
| from sentencepiece import sentencepiece_model_pb2 |
| from google.protobuf import struct_pb2 |
|
|
|
|
| @dataclasses.dataclass(frozen=True) |
| class TokensInfo: |
| token_ids: Sequence[int] |
| tokens: Sequence[bytes] |
| role: str = None |
|
|
|
|
| @dataclasses.dataclass(frozen=True) |
| class ComputeTokensResult: |
| """Represents token string pieces and ids output in compute_tokens function. |
| |
| Attributes: |
| tokens_info: Lists of tokens_info from the input. |
| The input `contents: ContentsType` could have |
| multiple string instances and each tokens_info |
| item represents each string instance. Each token |
| info consists tokens list, token_ids list and |
| a role. |
| """ |
|
|
| tokens_info: Sequence[TokensInfo] |
|
|
|
|
| class PreviewComputeTokensResult(ComputeTokensResult): |
| def token_info_list(self) -> Sequence[TokensInfo]: |
| import warnings |
|
|
| message = "PreviewComputeTokensResult.token_info_list is deprecated. Use ComputeTokensResult.tokens_info instead." |
| warnings.warn(message, DeprecationWarning, stacklevel=2) |
| return self.tokens_info |
|
|
|
|
| @dataclasses.dataclass(frozen=True) |
| class CountTokensResult: |
| """Represents an token numbers output in count_tokens function. |
| |
| Attributes: |
| total_tokens: number of total tokens. |
| """ |
|
|
| total_tokens: int |
|
|
|
|
| def _parse_hex_byte(token: str) -> int: |
| """Parses a hex byte string of the form '<0xXX>' and returns the integer value. |
| |
| Raises ValueError if the input is malformed or the byte value is invalid. |
| """ |
|
|
| if len(token) != 6: |
| raise ValueError(f"Invalid byte length: {token}") |
| if not token.startswith("<0x") or not token.endswith(">"): |
| raise ValueError(f"Invalid byte format: {token}") |
|
|
| try: |
| val = int(token[3:5], 16) |
| except ValueError: |
| raise ValueError(f"Invalid hex value: {token}") |
|
|
| if val >= 256: |
| raise ValueError(f"Byte value out of range: {token}") |
|
|
| return val |
|
|
|
|
| def _token_str_to_bytes( |
| token: str, type: sentencepiece_model_pb2.ModelProto.SentencePiece.Type |
| ) -> bytes: |
| if type == sentencepiece_model_pb2.ModelProto.SentencePiece.Type.BYTE: |
| return _parse_hex_byte(token).to_bytes(length=1, byteorder="big") |
| else: |
| return token.replace("▁", " ").encode("utf-8") |
|
|
|
|
| class _SentencePieceAdaptor: |
| r"""An internal tokenizer that can parse text input into tokens.""" |
|
|
| def __init__(self, tokenizer_name: str): |
| r"""Initializes the tokenizer. |
| |
| Args: |
| name: The name of the tokenizer. |
| """ |
| self._model_proto = load_model_proto(tokenizer_name) |
| self._tokenizer = get_sentencepiece(tokenizer_name) |
|
|
| def count_tokens(self, contents: Iterable[str]) -> CountTokensResult: |
| r"""Counts the number of tokens in the input.""" |
| tokens_list = self._tokenizer.encode(list(contents)) |
|
|
| return CountTokensResult( |
| total_tokens=sum(len(tokens) for tokens in tokens_list) |
| ) |
|
|
| def compute_tokens( |
| self, *, contents: Iterable[str], roles: Iterable[str] |
| ) -> ComputeTokensResult: |
| """Computes the tokens ids and string pieces in the input.""" |
| content_list = list(contents) |
| tokens_protos = self._tokenizer.EncodeAsImmutableProto(content_list) |
| roles = list(roles) |
|
|
| token_infos = [] |
| for tokens_proto, role in zip(tokens_protos, roles): |
| token_infos.append( |
| TokensInfo( |
| token_ids=[piece.id for piece in tokens_proto.pieces], |
| tokens=[ |
| _token_str_to_bytes( |
| piece.piece, self._model_proto.pieces[piece.id].type |
| ) |
| for piece in tokens_proto.pieces |
| ], |
| role=role, |
| ) |
| ) |
| return ComputeTokensResult(tokens_info=token_infos) |
|
|
|
|
| def _to_gapic_contents( |
| contents: ContentsType, |
| ) -> List[gapic_content_types.Content]: |
| """Converts a GenerativeModel compatible contents type to a gapic content.""" |
| _validate_contents_type_as_valid_sequence(contents) |
| _assert_no_image_contents_type(contents) |
| gapic_contents = _content_types_to_gapic_contents(contents) |
| |
| return gapic_contents |
|
|
|
|
| def _content_types_to_role_iterator(contents: ContentsType) -> Iterable[str]: |
| gapic_contents = _to_gapic_contents(contents) |
| |
| for content in gapic_contents: |
| for part in content.parts: |
| yield content.role |
|
|
|
|
| def _assert_no_image_contents_type(contents: ContentsType): |
| """Asserts that the contents type does not contain any image content.""" |
| if isinstance(contents, Image) or ( |
| isinstance(contents, Sequence) |
| and any(isinstance(content, Image) for content in contents) |
| ): |
| raise ValueError("Tokenizers do not support Image content type.") |
|
|
|
|
| def _is_string_inputs(contents: ContentsType) -> bool: |
| return ( |
| isinstance(contents, str) |
| or isinstance(contents, Sequence) |
| and all(isinstance(content, str) for content in contents) |
| ) |
|
|
|
|
| def _to_canonical_roles(contents: ContentsType) -> Iterable[str]: |
| if isinstance(contents, str): |
| yield "user" |
| elif isinstance(contents, Sequence) and all( |
| isinstance(content, str) for content in contents |
| ): |
| yield from ["user"] * len(contents) |
| else: |
| yield from _content_types_to_role_iterator(contents) |
|
|
|
|
| class _TextsAccumulator: |
| """Accumulates texts from contents and tools. |
| |
| This class is used to accumulate countable texts from contents and tools. |
| When user passes a unsupported fields that are added in the future, the new |
| fields might be only counted in remote tokenizer. In this case, the local |
| tokenizer should know that an unsupported new field exist in the content or |
| tool instances and raise error to avoid returning incorrect result to users. |
| |
| The machanism to detect unsupported fields introduced in the future: when |
| local tokenizer traversing the input instances, it is allowlist based text |
| accumulation. When a field is traversed and evaluated to be countable, the |
| value of this |
| field is copied to two places: (1) self._texts for inputs to sentencepiece |
| token count function. (2) a counted instance object in the recursive |
| function's return value. That's to say, after done current recurssion, |
| the instance(of same type as the input) only keeps the counted values. |
| If user sets unsupported future proto fields, they can be detected by |
| comparing the input instances equal to counted instances or not. |
| """ |
|
|
| def __init__(self): |
| self._texts = [] |
|
|
| def get_texts(self) -> Iterable[str]: |
| return self._texts |
|
|
| def add_texts(self, texts: Union[Iterable[str], str]) -> None: |
| if isinstance(texts, str): |
| self._texts.append(texts) |
| else: |
| self._texts.extend(texts) |
|
|
| def add_content(self, content: gapic_content_types.Content) -> None: |
| counted_content = gapic_content_types.Content() |
| for part in content.parts: |
| counted_part = gapic_content_types.Part() |
| if "file_data" in part or "inline_data" in part: |
| raise ValueError("Tokenizers do not support non-text content types.") |
| if "video_metadata" in part: |
| counted_part.video_metadata = part.video_metadata |
| if "function_call" in part: |
| self.add_function_call(part.function_call) |
| counted_part.function_call = part.function_call |
| if "function_response" in part: |
| self.add_function_response(part.function_response) |
| counted_part.function_response = part.function_response |
| if "text" in part: |
| counted_part.text = part.text |
| self._texts.append(part.text) |
| counted_content.parts.append(counted_part) |
| counted_content.role = content.role |
| if content._pb != counted_content._pb: |
| raise ValueError( |
| f"Content contains unsupported types for token counting. Supported fields {counted_content}. Got {content}." |
| ) |
|
|
| def add_function_call(self, function_call: gapic_tool_types.FunctionCall) -> None: |
| """Processes a function call and adds relevant text to the accumulator. |
| |
| Args: |
| function_call: The function call to process. |
| """ |
| self._texts.append(function_call.name) |
| counted_function_call = gapic_tool_types.FunctionCall(name=function_call.name) |
| counted_struct = self._struct_traverse(function_call._pb.args) |
| counted_function_call.args = counted_struct |
| if counted_function_call._pb != function_call._pb: |
| raise ValueError( |
| f"Function call argument contains unsupported types for token counting. Supported fields {counted_function_call}. Got {function_call}." |
| ) |
|
|
| def add_function_calls( |
| self, function_calls: Iterable[gapic_tool_types.FunctionCall] |
| ) -> None: |
| for function_call in function_calls: |
| self.add_function_call(function_call) |
|
|
| def add_tool(self, tool: gapic_tool_types.Tool) -> gapic_tool_types.Tool: |
| counted_tool = gapic_tool_types.Tool() |
| for function_declaration in tool.function_declarations: |
| counted_function_declaration = self._function_declaration_traverse( |
| function_declaration |
| ) |
| counted_tool.function_declarations.append(counted_function_declaration) |
| if counted_tool._pb != tool._pb: |
| raise ValueError( |
| f"Tool argument contains unsupported types for token counting. Supported fields {counted_tool}. Got {tool}." |
| ) |
|
|
| def add_tools(self, tools: Iterable[gapic_tool_types.Tool]) -> None: |
| for tool in tools: |
| self.add_tool(tool) |
|
|
| def add_function_responses( |
| self, function_responses: Iterable[gapic_tool_types.FunctionResponse] |
| ) -> None: |
| for function_response in function_responses: |
| self.add_function_response(function_response) |
|
|
| def add_function_response( |
| self, function_response: gapic_tool_types.FunctionResponse |
| ) -> None: |
| counted_function_response = gapic_tool_types.FunctionResponse() |
| self._texts.append(function_response.name) |
| counted_struct = self._struct_traverse(function_response._pb.response) |
| counted_function_response.name = function_response.name |
| counted_function_response.response = counted_struct |
| if counted_function_response._pb != function_response._pb: |
| raise ValueError( |
| f"Function response argument contains unsupported types for token counting. Supported fields {counted_function_response}. Got {function_response}." |
| ) |
|
|
| def _function_declaration_traverse( |
| self, function_declaration: gapic_tool_types.FunctionDeclaration |
| ) -> gapic_tool_types.FunctionDeclaration: |
| counted_function_declaration = gapic_tool_types.FunctionDeclaration() |
| self._texts.append(function_declaration.name) |
| counted_function_declaration.name = function_declaration.name |
| if function_declaration.description: |
| self._texts.append(function_declaration.description) |
| counted_function_declaration.description = function_declaration.description |
| if function_declaration.parameters: |
| counted_parameters = self._schema_traverse(function_declaration.parameters) |
| counted_function_declaration.parameters = counted_parameters |
| if function_declaration.response: |
| counted_response = self._schema_traverse(function_declaration.response) |
| counted_function_declaration.response = counted_response |
| return counted_function_declaration |
|
|
| def _schema_traverse(self, schema: openapi.Schema) -> openapi.Schema: |
| """Processes a schema and adds relevant text to the accumulator. |
| |
| Args: |
| schema: The schema to process. |
| |
| Returns: |
| The new schema object with only countable fields. |
| """ |
| counted_schema = openapi.Schema() |
| if "type_" in schema: |
| counted_schema.type = schema.type |
| if "title" in schema: |
| counted_schema.title = schema.title |
| if "default" in schema: |
| counted_schema.default = schema.default |
| if "format_" in schema: |
| self._texts.append(schema.format_) |
| counted_schema.format_ = schema.format_ |
| if "description" in schema: |
| self._texts.append(schema.description) |
| counted_schema.description = schema.description |
| if "enum" in schema: |
| self._texts.extend(schema.enum) |
| counted_schema.enum = schema.enum |
| if "required" in schema: |
| self._texts.extend(schema.required) |
| counted_schema.required = schema.required |
| if "property_ordering" in schema: |
| counted_schema.property_ordering = schema.property_ordering |
| if "items" in schema: |
| counted_schema_items = self._schema_traverse(schema.items) |
| counted_schema.items = counted_schema_items |
| if "properties" in schema: |
| d = {} |
| for key, value in schema.properties.items(): |
| self._texts.append(key) |
| counted_value = self._schema_traverse(value) |
| d[key] = counted_value |
| counted_schema.properties.update(d) |
| if "example" in schema: |
| counted_schema_example = self._value_traverse(schema._pb.example) |
| counted_schema.example = counted_schema_example |
| return counted_schema |
|
|
| def _struct_traverse(self, struct: struct_pb2.Struct) -> struct_pb2.Struct: |
| """Processes a struct and adds relevant text to the accumulator. |
| |
| Args: |
| struct: The struct to process. |
| |
| Returns: |
| The new struct object with only countable fields. |
| """ |
| counted_struct = struct_pb2.Struct() |
| self._texts.extend(list(struct.fields.keys())) |
| for key, val in struct.fields.items(): |
| counted_struct_fields = self._value_traverse(val) |
| if isinstance(counted_struct_fields, str): |
| counted_struct.fields[key] = counted_struct_fields |
| else: |
| counted_struct.fields[key].MergeFrom(counted_struct_fields) |
| return counted_struct |
|
|
| def _value_traverse(self, value: struct_pb2.Value) -> struct_pb2.Value: |
| """Processes a struct field and adds relevant text to the accumulator. |
| |
| Args: |
| struct: The struct field to process. |
| |
| Returns: |
| The new struct field object with only countable fields. |
| """ |
| kind = value.WhichOneof("kind") |
| counted_value = struct_pb2.Value() |
| if kind == "string_value": |
| self._texts.append(value.string_value) |
| counted_value.string_value = value.string_value |
| elif kind == "struct_value": |
| counted_struct = self._struct_traverse(value.struct_value) |
| counted_value.struct_value.MergeFrom(counted_struct) |
| elif kind == "list_value": |
| counted_list_value = struct_pb2.ListValue() |
| for item in value.list_value.values: |
| counted_value = self._value_traverse(item) |
| counted_list_value.values.append(counted_value) |
| counted_value.list_value.MergeFrom(counted_list_value) |
| return counted_value |
|
|
|
|
| class Tokenizer: |
| """A tokenizer that can parse text into tokens.""" |
|
|
| def __init__(self, tokenizer_name: str): |
| """Initializes the tokenizer. |
| |
| Do not use this constructor directly. Use get_tokenizer_for_model instead. |
| |
| Args: |
| name: The name of the tokenizer. |
| |
| """ |
| self._sentencepiece_adapter = _SentencePieceAdaptor(tokenizer_name) |
|
|
| def count_tokens( |
| self, |
| contents: ContentsType, |
| *, |
| tools: Optional[List["Tool"]] = None, |
| system_instruction: Optional[PartsType] = None, |
| ) -> CountTokensResult: |
| r"""Counts the number of tokens in the text-only contents. |
| |
| Args: |
| contents: The contents to count tokens for. |
| Supports either a list of Content objects (passing a multi-turn |
| conversation) or a value that can be converted to a single |
| Content object (passing a single message). |
| Supports |
| * str, Part, |
| * List[Union[str, Part]], |
| * List[Content] |
| Throws an error if the contents contain non-text content. |
| tools: A list of tools (functions) that the model can try calling. |
| system_instruction: The provided system instructions for the model. |
| Note: only text should be used in parts and content in each part |
| will be in a separate paragraph. |
| |
| Returns: |
| A CountTokensResult object containing the total number of tokens in |
| the contents. |
| """ |
|
|
| text_accumulator = _TextsAccumulator() |
| if _is_string_inputs(contents): |
| text_accumulator.add_texts(contents) |
| else: |
| gapic_contents = _to_gapic_contents(contents) |
| for content in gapic_contents: |
| text_accumulator.add_content(content) |
|
|
| if tools: |
| text_accumulator.add_tools((tool._raw_tool for tool in tools)) |
|
|
| if system_instruction: |
| if _is_string_inputs(system_instruction): |
| text_accumulator.add_texts(system_instruction) |
| else: |
| text_accumulator.add_content(_to_content(system_instruction)) |
|
|
| return self._sentencepiece_adapter.count_tokens(text_accumulator.get_texts()) |
|
|
| def compute_tokens(self, contents: ContentsType) -> ComputeTokensResult: |
| r"""Computes the tokens ids and string pieces in the text-only contents. |
| |
| Args: |
| contents: The contents to count tokens for. |
| Supports either a list of Content objects (passing a multi-turn |
| conversation) or a value that can be converted to a single |
| Content object (passing a single message). |
| Supports |
| * str, Part, |
| * List[Union[str, Part]], |
| * List[Content] |
| Throws an error if the contents contain non-text content. |
| |
| Returns: |
| A ComputeTokensResult object containing the tokens ids and string |
| pieces in the contents. |
| |
| Examples: |
| compute_tokens(["hello world", "what's the weather today"]) |
| outputs: |
| ComputeTokensResult(tokens_info=[TokensInfo(token_ids=[17534, 2134], tokens=[b'hello', b' world'], role='user'), TokensInfo(token_ids=[84264, 235341], tokens=[b'Goodbye', b'!'], role='user')], token_info_list=...The same as tokens_info) |
| |
| """ |
| text_accumulator = _TextsAccumulator() |
| if _is_string_inputs(contents): |
| text_accumulator.add_texts(contents) |
| else: |
| gapic_contents = _to_gapic_contents(contents) |
| for content in gapic_contents: |
| text_accumulator.add_content(content) |
|
|
| return self._sentencepiece_adapter.compute_tokens( |
| contents=text_accumulator.get_texts(), |
| roles=_to_canonical_roles(contents), |
| ) |
|
|
|
|
| class PreviewTokenizer(Tokenizer): |
| def compute_tokens(self, contents: ContentsType) -> PreviewComputeTokensResult: |
| return PreviewComputeTokensResult( |
| tokens_info=super().compute_tokens(contents).tokens_info |
| ) |
|
|
|
|
| def _get_tokenizer_for_model_preview(model_name: str) -> PreviewTokenizer: |
| """Returns a tokenizer for the given tokenizer name. |
| |
| Usage: |
| ``` |
| tokenizer = get_tokenizer_for_model("gemini-1.5-pro-001") |
| print(tokenizer.count_tokens("Hello world!")) |
| ``` |
| |
| Supported models can be found at |
| https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models. |
| |
| Args: |
| model_name: Specify the tokenizer is from which model. |
| """ |
| if not model_name: |
| raise ValueError("model_name must not be empty.") |
|
|
| return PreviewTokenizer(get_tokenizer_name(model_name)) |
|
|
|
|
| def get_tokenizer_for_model(model_name: str) -> Tokenizer: |
| """Returns a tokenizer for the given tokenizer name. |
| |
| Usage: |
| ``` |
| tokenizer = get_tokenizer_for_model("gemini-1.5-pro-001") |
| print(tokenizer.count_tokens("Hello world!")) |
| ``` |
| |
| Supported models can be found at |
| https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models. |
| |
| Args: |
| model_name: Specify the tokenizer is from which model. |
| """ |
| if not model_name: |
| raise ValueError("model_name must not be empty.") |
|
|
| return Tokenizer(get_tokenizer_name(model_name)) |
|
|