spoken language identification
Identify a spoken language using artificial intelligence (LID). The solution uses the convolutional neural network in order to detect language specific phonemes. It supports 3 languages: English, German and Spanish. The inspiration for the project came from the TopCoder contest, Spoken Languages 2.
Take a look at the Demo section to try the project yourself against real life content.
Dataset
New dataset was created from scratch.
LibriVox recordings were used to prepare the dataset. Particular attention was paid to a big variety of unique speakers. Big variance forces the network to concentrate more on language properties than a specific voice. Samples are equally balanced between languages, genders and speakers in order not to favour any subgroup. Finally speakers present in the test set, are not present in the train set. This helps estimate a generalization error.
More information at tomasz-oponowicz/spoken_language_dataset.
Architecture
The first step is to normalize input audio. Each sample is an FLAC audio file with:
- sample rate: 22050
- bit depth: 16
- channels: 1
- duration: 10 seconds (sharp)
Next filter banks are extracted from samples. Mean and variance normalization is applied. Then data is scaled with the Min/Max scaler.
Finally preprocessed data is passed to the convolutional neural network. Please notice the AveragePooling2D layer which improved the performance. This strategy is called global average pooling. It effectively forces the previous layers to produce the confidence maps.
The output is multiclass.
Performance
The score against the test set (out-of-sample) is 97% (F1 metric). Additionally the network generalizes well and presents high score against real life content, for example podcasts or TV news.
Sound effects or languages other than English, German or Spanish may be badly classified. If you want to work with noisy audio consider filtering noise out beforehand.
Demo
Prerequisites
- docker is installed (tested with 18.03.0)
Steps
-
Create a temporary directory and change the current directory:
$ mkdir examples && cd $_
-
Download samples:
NOTE: An audio file should contain speech and silence only. For example podcasts, interviews or audiobooks are a good fit. Sound effects or languages other than English, German or Spanish may be badly classified.
-
English (confidence 85.36%):
$ wget "https://javascriptair.podbean.com/mf/player-preload/nkdkps/048_JavaScript_Air_-_JavaScript_and_the_Web_Platform_The_Grand_Finale_.mp3" -O en.mp3
-
German (confidence 85.53%):
$ wget "http://mp3-download.ard.de/radio/radiofeature/auf-die-fresse-xa9c.l.mp3" -O de.mp3
-
Spanish (confidence 86.96%):
$ wget "http://mvod.lvlt.rtve.es/resources/TE_SCINCOC/mp3/2/8/1526585716282.mp3" -O es.mp3
...other examples of real life content are listed in the EXAMPLES.md.
-
-
Build the docker image:
$ docker build -t sli --rm https://github.com/tomasz-oponowicz/spoken_language_identification.git
-
Mount the
examples
directory and classify an audio file, for example:$ docker run --rm -it -v $(pwd):/data sli /data/en.mp3
...there are several options available through command line. For example you can tweak the noise reducer by increasing or decreasing the
silence-threshold
(0.5
by default):$ docker run --rm -it -v $(pwd):/data sli --silence-threshold=1 /data/es.mp3
Train
Prerequisites
- ffmpeg is installed (tested with 3.4.2)
- sox is installed (tested with 14.4.2)
- docker is installed (tested with 18.03.0)
Steps
-
Clone the repository:
$ git clone [email protected]:tomasz-oponowicz/spoken_language_identification.git
-
Go to the newly created directory:
$ cd spoken_language_identification
-
Generate samples:
-
Fetch the spoken_language_dataset dataset:
$ git submodule update --init --recursive
-
Go to the dataset directory:
$ cd spoken_language_dataset
-
Generate samples:
NOTE: Alternatively you can download the pregenerated dataset. Depending on your hardware it can save you 1-2 hours. After downloading, extract contents into
build/train
andbuild/test
directories.$ make build
-
Fix file permission of newly generated samples:
$ make fix_permissions
-
Return to the
spoken_language_identification
directory$ cd ..
-
-
Install dependencies
$ pip install -r requirements.txt
...the
tensorflow
package is installed by default (i.e. CPU support only). In order to speed up the training, install thetensorflow-gpu
package instead (i.e. GPU support). More information at Installing TensorFlow. -
Generate features from samples:
$ python features.py
-
Normalize features and build folds:
$ python folds.py
-
Train the model:
$ python model.py
...new model is stored at
model.h5
.
Release history
- 2018-07-06 / v1.0 / Initial version