| |
| import os |
| import shutil |
| import warnings |
| from types import SimpleNamespace |
| from unittest import mock |
|
|
| import pytest |
| from tokenizers import Tokenizer as HFTokenizer |
| from tokenizers.models import BPE |
| from transformers import AutoTokenizer |
| from transformers.utils import cached_file |
|
|
| import litgpt.config as config_module |
| from litgpt import PromptStyle, Tokenizer |
|
|
|
|
| |
| @pytest.mark.parametrize("config", config_module.configs, ids=[c["hf_config"]["name"] for c in config_module.configs]) |
| def test_tokenizer_against_hf(config, tmp_path): |
| config = config_module.Config(**config) |
|
|
| repo_id = f"{config.hf_config['org']}/{config.hf_config['name']}" |
| theirs = AutoTokenizer.from_pretrained(repo_id, token=os.getenv("HF_TOKEN")) |
|
|
| |
| hf_files = {} |
| for filename in ("tokenizer.json", "generation_config.json", "tokenizer.model", "tokenizer_config.json"): |
| try: |
| hf_file = cached_file(path_or_repo_id=repo_id, filename=filename) |
| hf_files[filename] = str(hf_file) |
| except Exception as ex: |
| warnings.warn(str(ex), RuntimeWarning) |
| if "tokenizer.json" not in hf_files and "tokenizer.model" not in hf_files: |
| raise ConnectionError("Unable to download any tokenizer files from HF") |
|
|
| |
| |
| tmp_path = tmp_path.rename(tmp_path.parent / config.hf_config["name"]) |
|
|
| for filename, hf_file in hf_files.items(): |
| shutil.copy(hf_file, str(tmp_path / filename)) |
|
|
| ours = Tokenizer(tmp_path) |
|
|
| assert ours.vocab_size == theirs.vocab_size |
| if config.name == "Mixtral-8x22B-v0.1": |
| pytest.xfail(reason="Mixtral certainly lists 32000 vocab in its config") |
| else: |
| assert ours.vocab_size == config.vocab_size |
|
|
| if config.name.startswith(("falcon", "stablecode", "Qwen2.5", "QwQ", "Qwen3")): |
| |
| assert isinstance(ours.bos_id, int) |
| assert theirs.bos_token_id is None |
| elif config.name.startswith("Falcon3"): |
| if isinstance(ours.bos_id, int): |
| assert theirs.bos_token_id is None |
| else: |
| assert ours.bos_id == theirs.bos_token_id is None |
| else: |
| assert ours.bos_id == theirs.bos_token_id |
|
|
| if config.name.startswith("stablecode"): |
| |
| assert ours.eos_id == 0 |
| assert ours.eos_id == theirs.eos_token_id or theirs.eos_token_id is None |
| else: |
| assert ours.eos_id == theirs.eos_token_id |
|
|
| prompt = "Hello, readers of this test!" |
| prompt = PromptStyle.from_config(config).apply(prompt) |
| actual = ours.encode(prompt) |
| expected = theirs.encode(prompt) |
| assert actual.tolist() == expected |
| assert ours.decode(actual) == theirs.decode(expected, skip_special_tokens=True) |
|
|
| if not config.name.startswith(("Mistral", "Mixtral")): |
| decoded_output = "".join([ours.decode(x) for x in actual]) |
| if ours.apply_decoding_fix and decoded_output[0] == " ": |
| decoded_output = decoded_output[1:] |
| assert decoded_output == ours.decode(actual), type(theirs) |
|
|
|
|
| def test_tokenizer_input_validation(): |
| with pytest.raises(NotADirectoryError, match="The checkpoint directory does not exist"): |
| Tokenizer("cocofruit") |
|
|
|
|
| @pytest.mark.parametrize("use_bos_by_default", (True, False)) |
| @pytest.mark.parametrize("encode_use_bos", (None, True, False)) |
| @pytest.mark.parametrize("encode_use_eos", (True, False)) |
| @pytest.mark.parametrize("processor_returns_bos", (True, False)) |
| @pytest.mark.parametrize("fake_return_ids", ([], [34, 8, 17, 2])) |
| def test_tokenizer_bos_eos( |
| tmp_path, use_bos_by_default, encode_use_bos, encode_use_eos, processor_returns_bos, fake_return_ids |
| ): |
| |
| HFTokenizer(BPE()).save(str(tmp_path / "tokenizer.json")) |
|
|
| tokenizer = Tokenizer(tmp_path) |
| tokenizer.bos_id = 0 |
| tokenizer.eos_id = 1 |
| tokenizer.use_bos = use_bos_by_default |
|
|
| if processor_returns_bos: |
| fake_return_ids = [tokenizer.bos_id] + fake_return_ids |
| fake_return_ids = SimpleNamespace(**dict(ids=fake_return_ids)) |
|
|
| with mock.patch.object(tokenizer.processor, "encode", return_value=fake_return_ids): |
| tokens = tokenizer.encode("Hello world", bos=encode_use_bos, eos=encode_use_eos).tolist() |
|
|
| if encode_use_bos or (encode_use_bos is None and use_bos_by_default): |
| assert tokens[0] == tokenizer.bos_id |
| else: |
| assert not tokens or tokens[0] != tokenizer.bos_id |
|
|
| if encode_use_eos: |
| assert tokens[-1] == tokenizer.eos_id |
| else: |
| assert not tokens or tokens[-1] != tokenizer.eos_id |
|
|
| |
| |
| assert max([id for id, token in enumerate(tokens) if token == tokenizer.bos_id], default=0) == 0 |
| assert max([id for id, token in enumerate(tokens[::-1]) if token == tokenizer.eos_id], default=0) == 0 |
|
|