This is a customized version of minishlab/potion-multilingual-128M. It has been quantized to int8 and stripped to optimize for our use case.
Description
- Compression: Quantized from FP32 to Int8.
- Vocabulary Stripping: Pruned from 500k+ tokens to 378k+ tokens (97MB), removing low-frequency non-CJK/Latin scripts.
- Performance: Maintains 99%+ semantic integrity for English and CJK scripts compared to the original model.
- Format: ONNX only.
- Downloads last month
- 43
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for alikia2x/potion-multilingual-128M-int8-strip
Base model
minishlab/potion-multilingual-128M