CSLM
Collection
Model weights and dataset for CSLM (Efficient Training for Cross-lingual Speech Language Models) β’ 3 items β’ Updated
CSLM-base is a base cross-lingual speech language model checkpoint from the ACL 2026 Findings project Efficient Training for Cross-lingual Speech Language Models.
It is designed for speech-to-speech / speech-conditioned generation workflows where speech units are used as model input.
Paper: https://arxiv.org/abs/2604.11096
If you use this model, please cite:
@misc{zhou2026efficienttrainingcrosslingualspeech,
title={Efficient Training for Cross-lingual Speech Language Models},
author={Yan Zhou and Qingkai Fang and Yun Hong and Yang Feng},
year={2026},
eprint={2604.11096},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2604.11096},
}
For questions, contact: zhouyan23z@ict.ac.cn