An efficient encoder-decoder architecture with top-down attention for speech separation
This repository is the official implementation of An efficient encoder-decoder architecture with top-down attention for speech separation Paper link.
@inproceedings{tdanet2023iclr,
title={An efficient encoder-decoder architecture with top-down attention for speech separation},
author={Li, Kai and Yang, Runxuan and Hu, Xiaolin},
booktitle={ICLR},
year={2023}
}
News
π₯ October, 2023: We have released the pre-trained models of our TDANet. You can download them from and .
π July, 2023: We are pleased to announce the update of our model training framework! This new framework has excellent versatility, and it can flexibly handle the training and testing tasks of various voice separation models.
Datasets
The LRS2 dataset contains thousands of video clips acquired through BBC. LRS2 contains a large amount of noise and reverberation interference, which is more challenging and closer to the actual environment than the WSJ0 and LibriSpeech corpora.
LRS2-2Mix is created by using the LRS2 corpus, where the training set, validation set and test set contain 20000, 5000 and 3000 utterances, respectively. The two different speaker audios from different scenes with 16 kHz sample rate were randomly selected from the LRS2 corpus and were mixed with signal-to-noise ratios sampled between -5 dB and 5 dB. The length of mixture audios is 2 seconds.
Dataset Download Link: Google Driver
Training and evaluation
Training
python DataPreProcess/process_librimix.py --in_dir=xxxx --out_dir=DataPreProcess/Libri2Mix
python audio_train.py --conf_dir=configs/tdanet.yml
Evaluation
python audio_test.py --conf_dir=Experiments/checkpoint/TDANet/conf.yml
Inference with Pretrained Model
import os
import torch
import look2hear.models
os.environ['CUDA_VISIBLE_DEVICES'] = "0"
model = look2hear.models.BaseModel.from_pretrain("JusperLee/TDANetBest-2ms-LRS2").cuda()
test_data = torch.randn(1, 1, 16000).cuda()
out = model(test_data)
print(out.shape)
Results
Our model achieves the following performance on :