| |
| import importlib |
| import re |
| from abc import abstractmethod |
| from json import dumps |
| from pathlib import Path |
| from typing import TYPE_CHECKING, Dict, List, Optional, Tuple, Type, Union |
|
|
| import yaml |
|
|
| from litgpt.config import Config |
|
|
| if TYPE_CHECKING: |
| from litgpt import Tokenizer |
|
|
|
|
| class PromptStyle: |
| """Base interface for prompt styles.""" |
|
|
| @abstractmethod |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return prompt |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ([tokenizer.eos_id],) |
|
|
| @classmethod |
| def from_name(cls, name: str) -> "PromptStyle": |
| return prompt_styles[name]() |
|
|
| @classmethod |
| def from_config(cls, config: Config) -> "PromptStyle": |
| return model_name_to_prompt_style(config.name) |
|
|
|
|
| class Default(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return prompt |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ([tokenizer.eos_id],) |
|
|
|
|
| class Alpaca(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| if kwargs.get("input"): |
| sys_prompt = sys_prompt or ( |
| "Below is an instruction that describes a task, paired with an input that provides further context. " |
| "Write a response that appropriately completes the request.\n\n" |
| ) |
| return f"{sys_prompt}### Instruction:\n{prompt}\n\n### Input:\n{kwargs['input']}\n\n### Response:\n" |
|
|
| sys_prompt = sys_prompt or ( |
| "Below is an instruction that describes a task. " |
| "Write a response that appropriately completes the request.\n\n" |
| ) |
| return f"{sys_prompt}### Instruction:\n{prompt}\n\n### Response:\n" |
|
|
|
|
| class FLAN(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or ( |
| "Below is an instruction that describes a task. " |
| "Write a response that appropriately completes the request.\n\n" |
| ) |
| return f"{sys_prompt}### Instruction:\n{prompt}\n\n### Response:\n" |
|
|
|
|
| class Longform(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or ( |
| "Below is an instruction that describes a task, paired with an input that provides further context. " |
| "Write a response that appropriately completes the request.\n\n" |
| ) |
| return f"{sys_prompt}### Instruction:\n{prompt}\n\n### Response:\n" |
|
|
|
|
| class StableLMAlpha(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or ( |
| "# StableLM Tuned (Alpha version)\n- StableLM is a helpful and harmless open-source AI language" |
| " model developed by StabilityAI.\n- StableLM is excited to be able to help the user, but will refuse to do" |
| " anything that could be considered harmful to the user.\n- StableLM is more than just an information" |
| " source, StableLM is also able to write poetry, short stories, and make jokes.\n- StableLM will refuse to" |
| " participate in anything that could harm a human." |
| ) |
| return f"<|SYSTEM|>{sys_prompt}<|USER|>{prompt}<|ASSISTANT|>" |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ( |
| [tokenizer.eos_id], |
| [tokenizer.token_to_id("<|SYSTEM|>")], |
| [tokenizer.token_to_id("<|ASSISTANT|>")], |
| [tokenizer.token_to_id("<|USER|>")], |
| ) |
|
|
|
|
| class StableLMZephyr(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"<|user|>\n{prompt}<|endoftext|>\n<|assistant|>\n" |
|
|
|
|
| class Falcon(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"{prompt}\nAnswer:" |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ( |
| [tokenizer.eos_id], |
| |
| |
| [tokenizer.token_to_id("User"), tokenizer.token_to_id(":")], |
| [193, tokenizer.token_to_id("User")], |
| ) |
|
|
|
|
| class Falcon3(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"<|user|>\n{prompt}<|endoftext|>\n<|assistant|>\n" |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ( |
| [tokenizer.eos_id], |
| [tokenizer.token_to_id("<|endoftext|>")], |
| ) |
|
|
|
|
| class Llama2FunctionCalling(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| |
| b_func, e_func = "<FUNCTIONS>", "</FUNCTIONS>\n\n" |
| b_inst, e_inst = "[INST]", "[/INST]" |
| b_sys, e_sys = "<<SYS>>\n", "\n<</SYS>>\n\n" |
| |
| function_metadata = { |
| "function": "search_bing", |
| "description": ( |
| "Search the web for content on Bing. This allows users to search online/the internet/the web for" |
| " content." |
| ), |
| "arguments": [{"name": "query", "type": "string", "description": "The search query string"}], |
| } |
|
|
| system_prompt = sys_prompt or ( |
| "You are a helpful, respectful and honest assistant. Always answer as helpfully as" |
| "possible. Your only response should be JSON formatted functions" |
| ) |
| |
| function_list = dumps(function_metadata).replace("{", "{{").replace("}", "}}") |
| return ( |
| f"{b_func}{function_list.strip()}{e_func}{b_inst}{b_sys}{system_prompt.strip()}{e_sys}{prompt}{e_inst}\n\n" |
| ) |
|
|
|
|
| class Llama2(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| b_inst, e_inst = "[INST]", "[/INST]" |
| b_sys, e_sys = "<<SYS>>\n", "\n<</SYS>>\n\n" |
| sys_prompt = sys_prompt or ( |
| "You are a helpful, respectful and honest assistant. Always answer as helpfully as" |
| " possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist," |
| " toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and" |
| " positive in nature.\n\nIf a question does not make any sense, or is not factually coherent, explain why" |
| " instead of answering something not correct. If you don't know the answer to a question, please don't" |
| " share false information." |
| ) |
| return f"{b_inst} {b_sys}{sys_prompt}{e_sys} {prompt} {e_inst} " |
|
|
|
|
| class Llama3(PromptStyle): |
| def apply( |
| self, prompt: Union[str, List[Dict[str, str]]], *, sys_prompt: Optional[str] = None, **kwargs: str |
| ) -> str: |
| default_system_prompt = sys_prompt or "You are a helpful assistant." |
|
|
| |
| if isinstance(prompt, str): |
| return ( |
| "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n" |
| f"{default_system_prompt}<|eot_id|>" |
| "<|start_header_id|>user<|end_header_id|>\n\n" |
| f"{prompt}<|eot_id|>" |
| "<|start_header_id|>assistant<|end_header_id|>\n\n" |
| ) |
| elif isinstance(prompt, list): |
|
|
| def encode_header(role: str) -> List[str]: |
| return [f"<|start_header_id|>{role}<|end_header_id|>\n\n"] |
|
|
| def encode_message(message: Dict[str, str]) -> List[str]: |
| tokens = encode_header(message["role"]) |
| |
| tokens.append(message["content"].strip()) |
| tokens.append("<|eot_id|>") |
| return tokens |
|
|
| def has_system_prompt(messages: List[Dict[str, str]]) -> bool: |
| return messages[0].get("role", "") == "system" if len(messages) else False |
|
|
| tokens = ["<|begin_of_text|>"] |
| if not has_system_prompt(prompt): |
| tokens.extend(encode_message({"role": "system", "content": default_system_prompt})) |
| for i, message in enumerate(prompt): |
| if i != 0 and message["role"] == "system": |
| raise ValueError("'system' role is only allowed at the beginning of the conversation list.") |
| if message["role"] not in ["assistant", "user", "system"]: |
| raise ValueError( |
| f"Unknown role: '{message['role']}'. Supported roles are 'assistant', 'user', and 'system'." |
| ) |
| tokens.extend(encode_message(message)) |
| tokens.extend(encode_header("assistant")) |
| return "".join(tokens) |
| else: |
| raise ValueError(f"Unsupported prompt type: {type(prompt)}") |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ( |
| [tokenizer.eos_id], |
| [tokenizer.token_to_id("<|eot_id|>")], |
| ) |
|
|
|
|
| class R1Base(PromptStyle): |
| def apply( |
| self, prompt: Union[str, List[Dict[str, str]]], *, sys_prompt: Optional[str] = None, **kwargs: str |
| ) -> str: |
| default_system_prompt = sys_prompt or "" |
|
|
| bos_token = "<|begin▁of▁sentence|>" |
| eos_token = "" |
|
|
| if isinstance(prompt, str): |
| return f"{default_system_prompt}<|User|>{prompt}<|Assistant|>" |
| elif isinstance(prompt, list): |
|
|
| def encode_message(message: Dict[str, str]) -> str: |
| role = message["role"] |
| content = message["content"].strip() |
|
|
| if role == "system": |
| return content |
| elif role == "user": |
| return f"<|User|>{content}" |
| elif role == "assistant": |
| return f"<|Assistant|>{content}{eos_token}" |
| else: |
| raise ValueError(f"Unknown role: '{role}'. Supported roles are 'assistant', 'user', and 'system'.") |
|
|
| |
| system_prompt = "" |
| if prompt[0].get("role") == "system": |
| system_prompt = prompt[0]["content"] |
| prompt = prompt[1:] |
|
|
| |
| formatted_prompt = system_prompt |
| for message in prompt: |
| formatted_prompt += encode_message(message) |
|
|
| formatted_prompt += "<|Assistant|>" |
| return formatted_prompt |
| else: |
| raise ValueError(f"Unsupported prompt type: {type(prompt)}") |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ( |
| [tokenizer.eos_id], |
| [tokenizer.token_to_id("<|end▁of▁sentence|>")], |
| ) |
|
|
|
|
| class FreeWilly2(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or "This is a system prompt, please behave and help the user." |
| return f"### System:\n{sys_prompt}\n\n### User:\n{prompt}\n\n### Assistant:\n" |
|
|
|
|
| class Platypus(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"### Instruction:\n\n{prompt}\n\n### Response:\n" |
|
|
|
|
| class StableCode(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"###Instruction\n{prompt}###Response\n" |
|
|
|
|
| class CodeLlama(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| |
| |
| |
| b_inst, e_inst = "[INST]", "[/INST]" |
| if sys_prompt: |
| b_sys, e_sys = "<<SYS>>\n", "\n<</SYS>>\n\n" |
| return f"{b_inst} {b_sys}{sys_prompt}{e_sys}{prompt} {e_inst}" |
| return f"{b_inst} {prompt} {e_inst}" |
|
|
|
|
| class Phi1(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"{prompt}\n\nAnswer:" |
|
|
| def stop_tokens(self, tokenizer: "Tokenizer") -> Tuple[List[int], ...]: |
| return ( |
| [tokenizer.eos_id], |
| [tokenizer.token_to_id("Answer"), tokenizer.token_to_id(":")], |
| [198, tokenizer.token_to_id("Answer"), tokenizer.token_to_id(":")], |
| |
| |
| |
| ) |
|
|
|
|
| class Phi2(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"Instruct: {prompt}\nOutput:" |
|
|
|
|
| class Phi3(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or "You are a helpful assistant." |
| return f"<|system|>\n{sys_prompt}<|end|>\n<|user|>\n{prompt}<|end|>\n<|assistant|>\n" |
|
|
|
|
| class Phi4(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| res = "" |
| if sys_prompt: |
| res += f"<|im_start|>system<|im_sep|>{sys_prompt}<|im_end|>" |
| res += f"<|im_start|>user<|im_sep|>{prompt}<|im_end|><|im_start|>assistant<|im_sep|>" |
| return res |
|
|
|
|
| class Phi4Reasoning(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = ( |
| sys_prompt |
| or "You are Phi, a language model trained by Microsoft to help users. Your role as an assistant involves thoroughly exploring questions through a systematic thinking process before providing the final precise and accurate solutions. This requires engaging in a comprehensive cycle of analysis, summarizing, exploration, reassessment, reflection, backtracing, and iteration to develop well-considered thinking process. Please structure your response into two main sections: Thought and Solution using the specified format: <think> {Thought section} </think> {Solution section}. In the Thought section, detail your reasoning process in steps. Each step should include detailed considerations such as analysing questions, summarizing relevant findings, brainstorming new ideas, verifying the accuracy of the current steps, refining any errors, and revisiting previous steps. In the Solution section, based on various attempts, explorations, and reflections from the Thought section, systematically present the final solution that you deem correct. The Solution section should be logical, accurate, and concise and detail necessary steps needed to reach the conclusion. Now, try to solve the following question through the above guidelines:" |
| ) |
| return f"<|im_start>system<|im_sep|>{sys_prompt}<|im_end|><|im_start|>user<|im_sep|>{prompt}<|im_end|><|im_start|>assistant<|im_sep|>" |
|
|
|
|
| class Phi4Mini(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| res = "" |
| if sys_prompt: |
| res += f"<|system|>{sys_prompt}<|end|>" |
| res += f"<|user|>{prompt}<|end|><|assistant|>" |
| return res |
|
|
|
|
| class Phi4MiniReasoning(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or "Your name is Phi, an AI math expert developed by Microsoft." |
| return f"<|system|>{sys_prompt}<|end|><|user|>{prompt}<|end|><|assistant|>" |
|
|
|
|
| class TinyLlama(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or "You are a friendly chatbot who always gives helpful, detailed, and polite answers." |
| return f"<|system|>\n{sys_prompt}</s>\n<|user|>\n{prompt}</s>\n<|assistant|>\n" |
|
|
|
|
| class Gemma(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model\n" |
|
|
|
|
| class OLMo(PromptStyle): |
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| return f"<|endoftext|><|user|>\n{prompt}\n<|assistant|>\n" |
|
|
|
|
| class ChatML(PromptStyle): |
| def __init__(self, system_message: Optional[str] = None): |
| self.system_message = system_message |
|
|
| def apply(self, prompt: str, *, sys_prompt: Optional[str] = None, **kwargs: str) -> str: |
| sys_prompt = sys_prompt or self.system_message |
| return ( |
| f"<|im_start|>system\n{sys_prompt}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant\n" |
| ) |
|
|
|
|
| class Qwen2_5(ChatML): |
| def __init__(self): |
| super().__init__("You are Qwen, created by Alibaba Cloud. You are a helpful assistant.") |
|
|
|
|
| class Qwen2_5_Math(ChatML): |
| def __init__(self): |
| super().__init__("Please reason step by step, and put your final answer within \\boxed{}.") |
|
|
|
|
| class QwQ(ChatML): |
| def __init__(self): |
| super().__init__( |
| "You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step." |
| ) |
|
|
|
|
| class Qwen3(ChatML): |
| def __init__(self): |
| super().__init__() |
|
|
|
|
| class SmolLM2(ChatML): |
| def __init__(self): |
| super().__init__("You are a helpful AI assistant named SmolLM, trained by Hugging Face") |
|
|
|
|
| class Salamandra(ChatML): |
| def __init__(self): |
| super().__init__( |
| "I am Salamandra, an AI language model developed at the Barcelona Supercomputing Centre (BSC) by the Language Technologies Unit. My knowledge base was last updated on August 2023. Today Date: 2024-09-30\nSoy Salamandra, un modelo lingüístico de IA desarrollado en el Barcelona Supercomputing Centre (BSC) por la Language Technologies Unit. Mi base de conocimientos se actualizó por última vez en agosto de 2023.\nSoc Salamandra, un model de llenguatge d'IA desenvolupat al Barcelona Supercomputing Centre (BSC) per la Language Technologies Unit." |
| ) |
|
|
|
|
| |
| prompt_styles: Dict[str, Type[PromptStyle]] = { |
| |
| "default": Default, |
| "alpaca": Alpaca, |
| "flan": FLAN, |
| "longform": Longform, |
| |
| "stablelm-alpha": StableLMAlpha, |
| "stablelm-zephyr": StableLMZephyr, |
| "falcon": Falcon, |
| "llama2-function-calling": Llama2FunctionCalling, |
| "llama2": Llama2, |
| "freewilly2": FreeWilly2, |
| "platypus": Platypus, |
| "stablecode": StableCode, |
| "codellama": CodeLlama, |
| "phi-1": Phi1, |
| "phi-2": Phi2, |
| "phi-3": Phi3, |
| "phi-4": Phi4, |
| "phi-4-reasoning": Phi4Reasoning, |
| "phi-4-mini": Phi4Mini, |
| "phi-4-mini-reasoning": Phi4MiniReasoning, |
| "tinyllama": TinyLlama, |
| "gemma": Gemma, |
| "llama3": Llama3, |
| "olmo": OLMo, |
| "qwen2.5": Qwen2_5, |
| "qwen2.5-math": Qwen2_5_Math, |
| "qwq": QwQ, |
| "qwen3": Qwen3, |
| "smollm2": SmolLM2, |
| "salamandra": Salamandra, |
| } |
|
|
|
|
| def model_name_to_prompt_style(model_name: str) -> PromptStyle: |
| if re.search(r"stablelm-tuned-alpha", model_name): |
| return StableLMAlpha() |
| if re.search(r"stablelm-zephyr-3b", model_name): |
| return StableLMZephyr() |
| if re.search("stablecode-instruct", model_name): |
| return StableCode() |
| if re.search(r"Falcon3.*-Instruct", model_name): |
| return Falcon3() |
| if re.search(r"falcon.*-instruct", model_name): |
| return Falcon() |
| if re.search("Llama-2-7b-chat-hf-function-calling-v2", model_name): |
| return Llama2FunctionCalling() |
| if re.search("Llama-2.*-chat", model_name): |
| return Llama2() |
| if re.search("Llama-3.*-Instruct", model_name): |
| return Llama3() |
| if re.search("Llama-3.*-Instruct-*", model_name): |
| return Llama3() |
| if re.search("OLMo-2.*-(Instruct|SFT|DPO)", model_name): |
| return Llama3() |
| if re.search("R1", model_name): |
| return R1Base() |
| if re.search("FreeWilly2", model_name): |
| return FreeWilly2() |
| if re.search("Platypus", model_name): |
| return Platypus() |
| if re.search("CodeLlama|Mi[sx]tral.*Instruct", model_name): |
| return CodeLlama() |
| if re.search("phi-1", model_name): |
| return Phi1() |
| if re.search("phi-2", model_name): |
| return Phi2() |
| if re.search("Phi-3", model_name): |
| return Phi3() |
| if re.search("Phi-4-reasoning", model_name): |
| return Phi4Reasoning() |
| if re.search("Phi-4-mini-reasoning", model_name): |
| return Phi4MiniReasoning() |
| if re.search("Phi-4-mini", model_name): |
| return Phi4Mini() |
| if re.search("phi-4", model_name): |
| return Phi4() |
| if re.search(r"tiny-llama.*chat", model_name): |
| return TinyLlama() |
| if re.search(r"(Code)?Gemma.*-it", model_name): |
| return Gemma() |
| if re.search(r"OLMo.*-hf", model_name): |
| return OLMo() |
| if re.search(r"Qwen2\.5-Math-.*", model_name): |
| return Qwen2_5_Math() |
| if re.search(r"Qwen2\.5-.*", model_name): |
| return Qwen2_5() |
| if re.search(r"QwQ-.*", model_name): |
| return QwQ() |
| if re.search(r"Qwen3-.*", model_name): |
| return Qwen3() |
| if re.search(r"SmolLM2.*-Instruct", model_name): |
| return SmolLM2() |
| if re.search(r"salamandra-.*-instruct", model_name): |
| return Salamandra() |
| return Default() |
|
|
|
|
| def save_prompt_style(style: Union[str, PromptStyle], checkpoint_dir: Path) -> None: |
| style = PromptStyle.from_name(style) if isinstance(style, str) else style |
| cls = type(style) |
| |
| config = {"class_path": f"{cls.__module__}.{cls.__name__}"} |
| with open(checkpoint_dir / "prompt_style.yaml", "w", encoding="utf-8") as file: |
| yaml.dump(config, file) |
|
|
|
|
| def load_prompt_style(checkpoint_dir: Path) -> PromptStyle: |
| with open(checkpoint_dir / "prompt_style.yaml", encoding="utf-8") as file: |
| config = yaml.safe_load(file) |
| |
| full_module_path, cls_name = config["class_path"].rsplit(".", 1) |
| module = importlib.import_module(full_module_path) |
| cls = getattr(module, cls_name) |
| return cls() |
|
|
|
|
| def has_prompt_style(checkpoint_dir: Path) -> bool: |
| return (checkpoint_dir / "prompt_style.yaml").is_file() |
|
|