Music Discriminator with LSTMs
- Discriminate music as being composed by humans or by a generative model.
- Use multiple layers of LSTM Recurrent Neural Nets
- Implementations in PyTorch and Keras.
Test trained LSTM model
In the ./weights/
you can find trained model weights and model architecture.
To test the model on your custom audio file, run
python3 predict_example.py path/to/custom/file.mp3
or to test the model on our custom files, run
python3 predict_example.py audio/classical_music.mp3
Audio features extracted
Dependencies
- Python3
- numpy
- librosa → for audio feature extraction
- Keras
pip install keras
- PyTorch
pip install torch torchvision
brew install libomp
Ideas for improving accuracy:
- Normalize MFCCs & other input features (Recurrent BatchNorm?)
- Decay learning rate
- How are we initing the weights?
- Better optimization hyperparameters (too little dropout)
- Do you have avoidable bias? How's your variance?