Tokenizer error?

#1
by 3ndetz - opened

image

looks like GGUF version is too bad...

but safetensors with patched tokenizers (from which I converted) was normal.
I made some patches to llama cpp hf gguf convert to make it work (replace sentencepiece for t5 with gpt2 tokenizer, and replace pre-toketnizer to gpt2), so it launched, but seems like there is trash in model outputs

Потом поищу решение, но пока заморожу на дня 2 или дольше. Если вдруг кто это читает (особенно, кто работал с фредом), буду рад помощи.
Вот скрипт в колабе для конверта.

Sign up or log in to comment