| |
|
|
| from typing import Optional |
|
|
| import pytest |
| import yaml |
|
|
| import litgpt.config |
| from litgpt import Config |
| from litgpt.prompts import ( |
| Alpaca, |
| Default, |
| Llama3, |
| Phi3, |
| PromptStyle, |
| has_prompt_style, |
| load_prompt_style, |
| prompt_styles, |
| save_prompt_style, |
| ) |
|
|
|
|
| def test_default_prompt_style(mock_tokenizer): |
| prompt_style = Default() |
| prompt = "This is a test prompt." |
| assert prompt_style.apply(prompt) == prompt |
| assert prompt_style.stop_tokens(mock_tokenizer) == ([mock_tokenizer.eos_id],) |
|
|
|
|
| @pytest.mark.parametrize("sys_prompt", [None, "You are a helpful coding assistant."]) |
| def test_sys_prompt(mock_tokenizer, sys_prompt: Optional[str]): |
| prompt_style = Phi3() |
| prompt = "This is a test prompt." |
| default_sys_prompt = "You are a helpful assistant." |
| response = f"<|system|>\n{sys_prompt or default_sys_prompt}<|end|>\n<|user|>\n{prompt}<|end|>\n<|assistant|>\n" |
| assert prompt_style.apply(prompt, sys_prompt=sys_prompt) == response |
| assert prompt_style.stop_tokens(mock_tokenizer) == ([mock_tokenizer.eos_id],) |
|
|
|
|
| @pytest.mark.parametrize("sys_prompt", [None, "You are a helpful coding assistant."]) |
| def test_sys_prompt_with_kwargs(mock_tokenizer, sys_prompt: Optional[str]): |
| prompt_style = Phi3() |
| prompt = "This is a test prompt." |
| default_sys_prompt = "You are a helpful assistant." |
| response = f"<|system|>\n{sys_prompt or default_sys_prompt}<|end|>\n<|user|>\n{prompt}<|end|>\n<|assistant|>\n" |
| assert prompt_style.apply(prompt, sys_prompt=sys_prompt, test=1) == response |
| assert prompt_style.stop_tokens(mock_tokenizer) == ([mock_tokenizer.eos_id],) |
|
|
|
|
| def test_prompt_style_from_name(): |
| for style_name in prompt_styles: |
| assert isinstance(PromptStyle.from_name(style_name), prompt_styles[style_name]) |
|
|
|
|
| def test_prompt_style_from_config(): |
| model_names = [ |
| "stablelm-tuned-alpha-3b", |
| "stablelm-tuned-alpha-7b", |
| "stablelm-zephyr-3b", |
| "stablecode-instruct-alpha-3b", |
| "falcon-7b-instruct", |
| "falcon-40b-instruct", |
| "Llama-2-7b-chat-hf", |
| "Llama-2-13b-chat-hf", |
| "Llama-2-70b-chat-hf", |
| "Llama-3-8B-Instruct", |
| "Llama-3-70B-Instruct", |
| "Llama-3.1-405B-Instruct", |
| "Gemma-2b-it", |
| "Gemma-7b-it", |
| "FreeWilly2", |
| "CodeLlama-7b-Instruct-hf", |
| "CodeLlama-13b-Instruct-hf", |
| "CodeLlama-34b-Instruct-hf", |
| "CodeLlama-70b-Instruct-hf", |
| "phi-1_5", |
| "phi-2", |
| "Phi-3-mini-4k-instruct", |
| "Mistral-7B-Instruct-v0.1", |
| "Mistral-7B-Instruct-v0.2", |
| "tiny-llama-1.1b-chat", |
| "Llama-2-7b-chat-hf-function-calling-v2", |
| ] |
|
|
| for c in litgpt.config.platypus: |
| model_names.append(c["name"]) |
|
|
| for model_name in model_names: |
| |
| assert not isinstance(PromptStyle.from_config(Config.from_name(model_name)), Default) |
|
|
|
|
| def test_apply_prompts(): |
| prompt = "Is a coconut a nut or a fruit?" |
| inp = "Optional input" |
|
|
| for style in prompt_styles.values(): |
| output = style().apply(prompt, input=inp) |
| assert prompt in output |
| if isinstance(style, Alpaca): |
| assert inp in output |
|
|
|
|
| class CustomPromptStyle(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs) -> str: |
| return prompt |
|
|
|
|
| def test_save_load_prompt_style(tmp_path): |
| |
| checkpoint_dir = tmp_path / "checkpoint" |
| checkpoint_dir.mkdir() |
| assert not has_prompt_style(checkpoint_dir) |
| save_prompt_style("alpaca", checkpoint_dir) |
| assert has_prompt_style(checkpoint_dir) |
| with open(checkpoint_dir / "prompt_style.yaml", encoding="utf-8") as file: |
| contents = yaml.safe_load(file) |
| assert contents == {"class_path": "litgpt.prompts.Alpaca"} |
| loaded = load_prompt_style(checkpoint_dir) |
| assert isinstance(loaded, Alpaca) |
|
|
| |
| checkpoint_dir = tmp_path / "custom" |
| checkpoint_dir.mkdir() |
| save_prompt_style(CustomPromptStyle(), checkpoint_dir) |
| with open(checkpoint_dir / "prompt_style.yaml", encoding="utf-8") as file: |
| contents = yaml.safe_load(file) |
| assert contents == {"class_path": "test_prompts.CustomPromptStyle"} |
| loaded = load_prompt_style(checkpoint_dir) |
| assert isinstance(loaded, CustomPromptStyle) |
|
|
|
|
| def test_multiturn_prompt(): |
| prompt = "What is the capital of France?" |
| msgs = [{"role": "user", "content": prompt}] |
| style = Llama3() |
| simple_output = style.apply(prompt) |
| multiturn_output = style.apply(msgs) |
| assert simple_output == multiturn_output |
|
|
| |
| msgs = [{"role": "system", "content": "You are not a helpful assistant."}, {"role": "user", "content": prompt}] |
| with_system_multiturn_output = style.apply(msgs) |
| assert "You are not a helpful assistant." in with_system_multiturn_output |
|
|
| |
| msgs = [ |
| {"role": "user", "content": prompt}, |
| ] |
| wo_system_multiturn_output = style.apply(msgs) |
| assert "You are a helpful assistant." in wo_system_multiturn_output |
|
|
| |
| msgs = [ |
| {"role": "system", "content": "You are a helpful AI assistant for travel tips and recommendations"}, |
| {"role": "user", "content": "What is France's capital?"}, |
| {"role": "assistant", "content": "Bonjour! The capital of France is Paris!"}, |
| {"role": "user", "content": "What can I do there?"}, |
| ] |
| multiturn_output = style.apply(msgs) |
|
|
| assert ( |
| multiturn_output |
| == """<|begin_of_text|><|start_header_id|>system<|end_header_id|> |
| |
| You are a helpful AI assistant for travel tips and recommendations<|eot_id|><|start_header_id|>user<|end_header_id|> |
| |
| What is France's capital?<|eot_id|><|start_header_id|>assistant<|end_header_id|> |
| |
| Bonjour! The capital of France is Paris!<|eot_id|><|start_header_id|>user<|end_header_id|> |
| |
| What can I do there?<|eot_id|><|start_header_id|>assistant<|end_header_id|> |
| |
| """ |
| ) |
|
|
| |
| msgs = [ |
| {"role": "user", "content": "What is France's capital?"}, |
| {"role": "assistant", "content": "Bonjour! The capital of France is Paris!"}, |
| {"role": "user", "content": "What can I do there?"}, |
| ] |
| multiturn_output = style.apply(msgs) |
|
|
| assert ( |
| multiturn_output |
| == """<|begin_of_text|><|start_header_id|>system<|end_header_id|> |
| |
| You are a helpful assistant.<|eot_id|><|start_header_id|>user<|end_header_id|> |
| |
| What is France's capital?<|eot_id|><|start_header_id|>assistant<|end_header_id|> |
| |
| Bonjour! The capital of France is Paris!<|eot_id|><|start_header_id|>user<|end_header_id|> |
| |
| What can I do there?<|eot_id|><|start_header_id|>assistant<|end_header_id|> |
| |
| """ |
| ) |
|
|
| |
| content = "this is {random} {system} {user}" |
| msgs = [{"role": "user", "content": content}] |
| output = style.apply(msgs) |
| simple_output = style.apply(content) |
| assert output == simple_output |
|
|