/YorNoise-for-MCLNN

Primary LanguageBatchfileMIT LicenseMIT

license

YorNoise dataset for MCLNN

The YorNoise environmental sound dataset.

Clip Duration Format Count Categories
4 secs .wav 1527 2

Dataset Summary:

  • Clips are 4-seconds in length with 44100 Hz sampling rates.
  • The dataset is released into predefined 10-fold splits for cross-validation.
  • It is combined through the experiments with the UrbanSound8k dataset to form a 12-classes dataset.

This folder contains:

  • Scripts required to prepare the UrbanSound8k dataset for the MCLNN processing.

  • Pretrained weights and indices for the 10-fold cross-validation in addition to the standardization parameters to replicate the results in:

    Fady Medhat, David Chesmore and John Robinson, "Recognition of Acoustic Events Using Masked Conditional Neural Networks," 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA)

Prepossessing

The preprocessing involved in preparing the YorNoise dataset is resampling to 22050 Hz.

Preparation scripts prerequisites

The preparation scripts require the following packages to be installed beforehand:

  • ffmpeg version N-81489-ga37e6dd
  • numpy 1.11.2+mkl
  • librosa 0.4.0
  • h5py 2.6.0yornoise_storage_ordering.txt

Steps

  1. Download the UrbanSound8k and execute its preprocessing scripts in the preparation scripts following the order of their labels.
  2. Download the YorNoise dataset and execute the scripts in the preparation scripts following the order of their labels.
  3. Make sure the files are ordered following the YorNoise_storage_ordering file.
  4. Configure the spectrogram transformation within the Dataset Transformer and generate the MCLNN-Ready hdf5 for the dataset using the YorNoiseUrbansound8k_MCLNN.csv file.
  5. Generate the indices for the folds using the Index Generator script.