[go: nahoru, domu]

Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Could not find results for multi-speaker audio files #2

Open
akshayaCap opened this issue Sep 5, 2018 · 5 comments
Open

Could not find results for multi-speaker audio files #2

akshayaCap opened this issue Sep 5, 2018 · 5 comments

Comments

@akshayaCap
Copy link

Hi,
I was going through your repository. I could not find results of LSTM and BLSTM on the 2 speaker .wav (audio files) generated by you. Can you please add them?

Also, have you tried this algorithm with multiple speaker with added noise? If yes, can you share the results?

@aishoot
Copy link
Owner
aishoot commented Sep 7, 2018

@akshayaCap Hello, I have uploaded .wav results of 2 speakers - "6-separated_result_BLSTM" and "7-separated_result_LSTM". As you say "with multiple speaker with added noise", one speaker can be regarded as target speaker while other speakers can be viewed as noise. And the algorithms are in first two folders.

@akshayaCap
Copy link
Author

@pchao6 thanks for your reply. Input files are missing in these folders. Can you please add them?

@aishoot
Copy link
Owner
aishoot commented Sep 9, 2018

@akshayaCap I'm sorry for sharing the input files. The input dataset WSJ0 needs paid license. There's the WSJ0 corpus website where you can purchase.

@akshayaCap
Copy link
Author

@pchao6 thank-you for clarification.

  1. Can you please share results on vctk-corpus as it is a freely available dataset.
  2. Also, is it possible for you to share script for inference on any noisy .wav file for a particular sampling rate.

@aishoot
Copy link
Owner
aishoot commented Sep 11, 2018

@akshayaCap Thanks for your interest in my work. Fisrtly, I haven't done a separation experiment on the VCTK dataset, but you can try it. Secondly, you can just replace one of the two speakers' .wav files with noise data when creating a mixed data set. Other experiment settings, such as codes, are the same. You can try it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants