Update README.md
Browse files
README.md
CHANGED
|
@@ -7,5 +7,8 @@ base_model:
|
|
| 7 |
|
| 8 |
This only works with the token ID directly. The tokenizer is completely busted.
|
| 9 |
|
|
|
|
|
|
|
|
|
|
| 10 |
CosyVoice also has a rich pre- and post- processing on top of the LLM step, so you can't do TTS out of the box with llamacpp.
|
| 11 |
Nevertheless, the LLM step is the slowest, and switching from pytorch to llamacpp yields 10x perf gain.
|
|
|
|
| 7 |
|
| 8 |
This only works with the token ID directly. The tokenizer is completely busted.
|
| 9 |
|
| 10 |
+
This is also missing the bias tensor on the decoding head. I don't think there's a way to do it without forking llamacpp.
|
| 11 |
+
I have not benchmarked the impact on quality due to the missing bias tensor.
|
| 12 |
+
|
| 13 |
CosyVoice also has a rich pre- and post- processing on top of the LLM step, so you can't do TTS out of the box with llamacpp.
|
| 14 |
Nevertheless, the LLM step is the slowest, and switching from pytorch to llamacpp yields 10x perf gain.
|