/Birdsong_separation

Unsupervised birdsong separation for classification

Primary LanguageJupyter NotebookMIT LicenseMIT

Sound Separation of Bird Calls Using MixIT

Birds are all around us, and just by listening, we can learn many things about our environment. Ecologists use birds to understand food systems and forest health. For example, if there are more woodpeckers in a forest, that means there’s a lot of dead wood.

The massive amount of available field recordings of birds presents an opportunity to use machine learning (ML) to automatically track bird populations and better understand ecosystems.

Google researchers have developped an unsupervised method called mixture invariant training (MixIT) to separate birdsongs and improve species classification. Below is a demonstration of birdsong separation from High Sierras (the video show the mel-spectogram of the mixed audio and highlight the audio separated into different tracks) :

birbsep_aiblog_caples.mp4

References

[1] Scott Wisdom, Efthymios Tzinis, Hakan Erdogan, Ron J. Weiss, Kevin Wilson, John R. Hershey, "Unsupervised Sound Separation Using Mixture Invariant Training", Advances in Neural Information Processing Systems, 2020.

[2] Tom Denton, Scott Wisdom, John R. Hershey, "Improving Bird Classification with Unsupervised Sound Separation", Proc. IEEE International Conference on Audio, Speech, and Signal Processing (ICASSP), 2022.

License

Copyright © 2023 Data Saiyentist.
This project is MIT License licensed.