Tinysoft commited on
Commit
d3e9f67
·
verified ·
1 Parent(s): b5feafe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -7,5 +7,8 @@ base_model:
7
 
8
  This only works with the token ID directly. The tokenizer is completely busted.
9
 
 
 
 
10
  CosyVoice also has a rich pre- and post- processing on top of the LLM step, so you can't do TTS out of the box with llamacpp.
11
  Nevertheless, the LLM step is the slowest, and switching from pytorch to llamacpp yields 10x perf gain.
 
7
 
8
  This only works with the token ID directly. The tokenizer is completely busted.
9
 
10
+ This is also missing the bias tensor on the decoding head. I don't think there's a way to do it without forking llamacpp.
11
+ I have not benchmarked the impact on quality due to the missing bias tensor.
12
+
13
  CosyVoice also has a rich pre- and post- processing on top of the LLM step, so you can't do TTS out of the box with llamacpp.
14
  Nevertheless, the LLM step is the slowest, and switching from pytorch to llamacpp yields 10x perf gain.