| #pragma once |
|
|
| #include "llama.h" |
|
|
| #include <string> |
| #include <vector> |
| #include <memory> |
| #include <unordered_map> |
|
|
| struct LLM_KV; |
| struct llama_model_loader; |
|
|
| struct llama_vocab { |
| struct token_data { |
| std::string text; |
| float score; |
| llama_token_attr attr; |
| }; |
|
|
| llama_vocab(); |
| ~llama_vocab(); |
|
|
| void load(llama_model_loader & ml, const LLM_KV & kv); |
|
|
| enum llama_vocab_type get_type() const; |
| enum llama_vocab_pre_type get_pre_type() const; |
|
|
| uint32_t n_tokens() const; |
| uint32_t n_token_types() const; |
|
|
| std::string type_name() const; |
|
|
| bool is_normal (llama_token id) const; |
| bool is_unknown (llama_token id) const; |
| bool is_control (llama_token id) const; |
| bool is_byte (llama_token id) const; |
| bool is_user_defined(llama_token id) const; |
| bool is_unused (llama_token id) const; |
| bool is_eog (llama_token id) const; |
|
|
| uint8_t token_to_byte(llama_token id) const; |
| llama_token byte_to_token(uint8_t ch) const; |
|
|
| llama_token text_to_token(const std::string & text) const; |
|
|
| const token_data & get_token_data(llama_token id) const; |
|
|
| const char * token_get_text (llama_token id) const; |
| float token_get_score(llama_token id) const; |
| llama_token_attr token_get_attr (llama_token id) const; |
|
|
| llama_token token_bos() const; |
| llama_token token_eos() const; |
| llama_token token_eot() const; |
| llama_token token_eom() const; |
| llama_token token_unk() const; |
| llama_token token_sep() const; |
| llama_token token_nl () const; |
| llama_token token_pad() const; |
|
|
| llama_token token_prefix() const; |
| llama_token token_middle() const; |
| llama_token token_suffix() const; |
|
|
| llama_token token_fim_pre() const; |
| llama_token token_fim_suf() const; |
| llama_token token_fim_mid() const; |
| llama_token token_fim_pad() const; |
| llama_token token_fim_rep() const; |
| llama_token token_fim_sep() const; |
|
|
| bool get_add_space_prefix () const; |
| bool get_add_bos () const; |
| bool get_add_eos () const; |
| bool get_ignore_merges () const; |
| bool get_clean_spaces () const; |
| bool get_remove_extra_whitespaces () const; |
| bool get_escape_whitespaces () const; |
| bool get_treat_whitespace_as_suffix() const; |
|
|
| int max_token_len() const; |
|
|
| int find_bpe_rank(const std::string & token_left, const std::string & token_right) const; |
|
|
| int32_t tokenize( |
| const char * text, |
| int32_t text_len, |
| llama_token * tokens, |
| int32_t n_tokens_max, |
| bool add_special, |
| bool parse_special) const; |
|
|
| std::vector<llama_token> tokenize( |
| const std::string & raw_text, |
| bool add_special, |
| bool parse_special = false) const; |
|
|
| |
| int32_t token_to_piece( |
| llama_token token, |
| char * buf, |
| int32_t length, |
| int32_t lstrip, |
| bool special) const; |
|
|
| |
| const std::string & token_to_piece(llama_token token) const; |
|
|
| int32_t detokenize( |
| const llama_token * tokens, |
| int32_t n_tokens, |
| char * text, |
| int32_t text_len_max, |
| bool remove_special, |
| bool unparse_special) const; |
|
|
| std::string detokenize( |
| const std::vector<llama_token> & tokens, |
| bool special) const; |
|
|
| void print_info() const; |
|
|
| void set_eos_bos(llama_token eos, llama_token bos); |
| const std::unordered_map<std::string, llama_token> & get_token_to_id() const; |
| const std::vector<token_data> & get_id_to_token() const; |
|
|
| private: |
| struct impl; |
| std::unique_ptr<impl> pimpl; |
| }; |
|
|