/UltraDualPathCompression

A Pytorch-based implementation of the compression and decompression module in "Ultra Dual-Path Compression For Joint Echo Cancellation And Noise Suppression".

Primary LanguageJupyter NotebookApache License 2.0Apache-2.0

Ultra Dual-Path Compression and Decompression

This is the repository for a Pytorch-based implementation of the compression and decompression module in "Ultra Dual-Path Compression For Joint Echo Cancellation And Noise Suppression". The ultra dual-path compression module can compress the input multi-track spectra with large numbers of frames and frequency (T-F) bins into feature maps with small numbers of T-F bins, facilitating the fast processing for dual-path models (e.g., fullsubnet, 2D-convolution network). The decompression module transforms the compressed feature map back to the shapes of spectra for further processing.

The latest codes are recommended to be found in ultra_dual_path_compression.ipynb, including dual-path compression, PostNet, an example of a dual-path GRU module, an example of a whole front-end network, and some examples of usage. Note that the codes of the dual-path GRU module and the whole front-end network are only for demonstration purposes and differ from what is in the article. Due to policy restrictions, the whole front-end network in the article will not be open-sourced at this time. ultra_dual_path_compression.py contains some legacy code, which will not be updated in the future.

Demos can be found in DemoPage.

Please refer to our paper with the latest version on Arxiv for details. This paper is also accepted by INTERSPEECH2023.

Please cite the paper if you found this module useful.

@article{DBLP:journals/corr/abs-2308-11053,
  author       = {Hangting Chen and
                  Jianwei Yu and
                  Yi Luo and
                  Rongzhi Gu and
                  Weihua Li and
                  Zhuocheng Lu and
                  Chao Weng},
  title        = {Ultra Dual-Path Compression For Joint Echo Cancellation And Noise
                  Suppression},
  journal      = {CoRR},
  volume       = {abs/2308.11053},
  year         = {2023},
  url          = {https://doi.org/10.48550/arXiv.2308.11053},
  doi          = {10.48550/arXiv.2308.11053},
  eprinttype    = {arXiv},
  eprint       = {2308.11053},
  timestamp    = {Fri, 25 Aug 2023 12:09:57 +0200},
  biburl       = {https://dblp.org/rec/journals/corr/abs-2308-11053.bib},
  bibsource    = {dblp computer science bibliography, https://dblp.org}
}

Disclaimer

This is not an officially supported Tencent product.