Pinned Repositories
BeamformIt
BeamformIt acoustic beamforming software
cheatsheets
Official Matplotlib cheat sheets
Dive-into-DL-PyTorch
本项目将《动手学深度学习》(Dive into Deep Learning)原书中的MXNet实现改为PyTorch实现。
DNS-Challenge
This repo contains the scripts, models, and required files for the ICASSP 2021 Deep Noise Suppression (DNS) Challenge.
examples
A set of examples around pytorch in Vision, Text, Reinforcement Learning, etc.
pb_chime5
Speech enhancement system for the CHiME-5 dinner party scenario
percepnet
percepnet implemented using Keras, still need to be optimized and tuned.
pyroomacoustics
Pyroomacoustics is a package for audio signal processing for indoor applications. It was developed as a fast prototyping platform for beamforming algorithms in indoor scenarios.
pytorch-tutorial
PyTorch Tutorial for Deep Learning Researchers
rnnoise
Recurrent neural network for audio noise reduction
AmosCch's Repositories
AmosCch/percepnet
percepnet implemented using Keras, still need to be optimized and tuned.
AmosCch/BeamformIt
BeamformIt acoustic beamforming software
AmosCch/cheatsheets
Official Matplotlib cheat sheets
AmosCch/Dive-into-DL-PyTorch
本项目将《动手学深度学习》(Dive into Deep Learning)原书中的MXNet实现改为PyTorch实现。
AmosCch/DNS-Challenge
This repo contains the scripts, models, and required files for the ICASSP 2021 Deep Noise Suppression (DNS) Challenge.
AmosCch/examples
A set of examples around pytorch in Vision, Text, Reinforcement Learning, etc.
AmosCch/pb_chime5
Speech enhancement system for the CHiME-5 dinner party scenario
AmosCch/pyroomacoustics
Pyroomacoustics is a package for audio signal processing for indoor applications. It was developed as a fast prototyping platform for beamforming algorithms in indoor scenarios.
AmosCch/pytorch-tutorial
PyTorch Tutorial for Deep Learning Researchers
AmosCch/rnnoise
Recurrent neural network for audio noise reduction
AmosCch/svoice
We provide a PyTorch implementation of the paper Voice Separation with an Unknown Number of Multiple Speakers In which, we present a new method for separating a mixed audio sequence, in which multiple voices speak simultaneously. The new method employs gated neural networks that are trained to separate the voices at multiple processing steps, while maintaining the speaker in each output channel fixed. A different model is trained for every number of possible speakers, and the model with the largest number of speakers is employed to select the actual number of speakers in a given sample. Our method greatly outperforms the current state of the art, which, as we show, is not competitive for more than two speakers.
AmosCch/TAPLoss
AmosCch/webrtc
WebRTC sub-repo dependency for WebRTC SDK
AmosCch/WebRTC-3A1V
AEC, AGC, ANS, VAD, CNG in WebRTC