rVAD
Description
Matlab and Python libraries for an unsupervised method for robust voice activity detection (rVAD) or speech activity detection (SAD), as presented in rVAD: An Unsupervised Segment-Based Robust Voice Activity Detection Method, Computer Speech & Language, 2020 or its arXiv version.
The rVAD paper published in Computer Speech & Language won International Speech Communication Association (ISCA) 2022 Best Research Paper Award.
The rVAD method consists of two passes of denoising followed by a VAD stage. It has been applied as a preprocessor for a wide range of applications, such as speech recognition, speaker identification, language identification, age and gender identification, self-supervised learning, human-robot interaction, audio archive segmentation, and so on as in Google Scholar.
The method is unsupervised to make it applicable to a broad range of acoustic environments, and it is optimized considering both noisy and clean conditions.
The rVAD (out of the box) ranks the 4th place (out of 27 supervised/unsupervised systems) in a Fearless Steps Speech Activity Detection Challenge.
The rVAD paper is among the most cited articles from Computer Speech and Language published since 2018 (the 6th place), in 2022 and 2023.
Source code for rVAD:
Source code in Matlab for rVAD (including both rVAD and rVAD-fast) is available under the rVAD2.0 folder. It is straightforward to use: Simply call the function vad.m. Some Matlab functions and their modified versions from the publicly available VoiceBox are included with kind permission of Mike Brookes.
Source code in Python for rVAD-fast is available under the rVADfast_py_2.0 folder. Source code for rVAD-fast to take streaming audio in is included too.
rVAD-fast is 10+ times faster than rVAD while rVAD has superior performance.
The rVADfast library is available as a python package installable via: pip install rVADfast. See rVADfast GitHub page for more details.
Reference VAD for Aurora 2 database:
The frame-by-frame reference VAD was generated from the clean set of Aurora 2 using forced-alignment speech recognition and has been used as a 'ground truth' for evaluating VAD algorithms. Our study shows that forced-alignment ASR performs as well as a human expert labeler for generating VAD references, as detailed in Comparison of Forced-Alignment Speech Recognition and Humans for Generating Reference VAD. Here are the generated reference VAD for the training set and the reference VAD for the test set.