How make inference on a single WAV file using pre-train model?
faranaziz opened this issue · 3 comments
No can find example for a single file.
Thanks you.
Hi
It has been awhile since I used it, but after you download the repo and install it as mentioned in https://github.com/haoxiangsnr/FullSubNet/blob/main/docs/prerequisites.md
, you just follow the instructions on https://github.com/haoxiangsnr/FullSubNet/blob/main/docs/getting_started.md
for Inference and Applying a Pre-trained Model .
as I remember you
- download the checkpoint model
fullsubnet_best_model_58epochs.tar
fromhttps://github.com/haoxiangsnr/FullSubNet/releases
and put it in a folder you create (I'll refer to the path to this folder as<model>
) - create a folder for your wave file (I'll refer to the path to this folder as
<input>
) and put your wave file in it - add the
<input>
path to the\recipes\dns_interspeech_2020\fullsubnet\inference.toml
file as described in step 1 under Inference mentioned above
...
[dataset.args]
dataset_dir_list = [
"<input>"
]
...
- create a folder for your output (I'll refer to the path to this folder as
<output>
) - follow step 2 under Inference mentioned above
cd FullSubNet/recipes/dns_interspeech_2020
# One GPU is used by default
python inference.py \
-C fullsubnet/inference.toml \
-M <model> \
-O <output>
Your denoised file will be in <output>
Good Luck!
Hi
It has been awhile since I used it, but after you download the repo and install it as mentioned inhttps://github.com/haoxiangsnr/FullSubNet/blob/main/docs/prerequisites.md
, you just follow the instructions onhttps://github.com/haoxiangsnr/FullSubNet/blob/main/docs/getting_started.md
for Inference and Applying a Pre-trained Model .as I remember you
- download the checkpoint model
fullsubnet_best_model_58epochs.tar
fromhttps://github.com/haoxiangsnr/FullSubNet/releases
and put it in a folder you create (I'll refer to the path to this folder as<model>
)- create a folder for your wave file (I'll refer to the path to this folder as
<input>
) and put your wave file in it- add the
<input>
path to the\recipes\dns_interspeech_2020\fullsubnet\inference.toml
file as described in step 1 under Inference mentioned above... [dataset.args] dataset_dir_list = [ "<input>" ] ...
- create a folder for your output (I'll refer to the path to this folder as
<output>
)- follow step 2 under Inference mentioned above
cd FullSubNet/recipes/dns_interspeech_2020 # One GPU is used by default python inference.py \ -C fullsubnet/inference.toml \ -M <model> \ -O <output>
Your denoised file will be in
<output>
Good Luck!
@sjscotti Great, thanks for your attention and help.
No can find example for a single file.
Thanks you.
Hi,
Thanks for your attention.
@sjscotti has given you a lot of details about how to enhance one file.
If you have any other questions, please feel free to contact me.