Papers
arxiv:2010.13154

Attention is All You Need in Speech Separation

Published on Oct 25, 2020
Authors:
,
,
,
,

Abstract

SepFormer, a Transformer-based model, achieves state-of-the-art speech separation performance on WSJ0-2/3mix datasets with parallelization and efficiency advantages.

AI-generated summary

Recurrent Neural Networks (RNNs) have long been the dominant architecture in sequence-to-sequence learning. RNNs, however, are inherently sequential models that do not allow parallelization of their computations. Transformers are emerging as a natural alternative to standard RNNs, replacing recurrent computations with a multi-head attention mechanism. In this paper, we propose the SepFormer, a novel RNN-free Transformer-based neural network for speech separation. The SepFormer learns short and long-term dependencies with a multi-scale approach that employs transformers. The proposed model achieves state-of-the-art (SOTA) performance on the standard WSJ0-2/3mix datasets. It reaches an SI-SNRi of 22.3 dB on WSJ0-2mix and an SI-SNRi of 19.5 dB on WSJ0-3mix. The SepFormer inherits the parallelization advantages of Transformers and achieves a competitive performance even when downsampling the encoded representation by a factor of 8. It is thus significantly faster and it is less memory-demanding than the latest speech separation systems with comparable performance.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2010.13154
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 21

Browse 21 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2010.13154 in a dataset README.md to link it from this page.

Spaces citing this paper 65

Collections including this paper 1