How to convert audio data into test data of algorithm for testing
zyc1310517843 opened this issue · 1 comments
Describe the question
A clear and concise description of what the question is.
My background
Have I read the README.md
file?
- yes
Have I searched for similar questions from closed issues?
- yes
Have I tried to find the answers in the paper Fully Supervised Speaker Diarization?
- yes
Hello, I have read README.md I want to convert my audio data into the test and training data needed by the model. How can I do this? I have also tried the third-party methods provided in README.md. However, they are for specific data sets, such as TIMIT, using our own audio data can not run successfully. Thank you very much for your guidance.
You can use any continuous speaker embedding technique to do that.
I listed a few here: https://github.com/wq2012/awesome-diarization#speaker-embedding
Not all of them supports continuous embeddings. You need to do some research on your own here.