Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Have you overcome overfitting problem] #3

Open
vuthede opened this issue Oct 5, 2020 · 8 comments
Open

[Have you overcome overfitting problem] #3

vuthede opened this issue Oct 5, 2020 · 8 comments

Comments

@vuthede
Copy link

vuthede commented Oct 5, 2020

Hi there,
@vitrioil Just want to ask have you overcame the overfitting problem that you reported in README?
Thanks, Do you have any idea of your overfitting? and any idea to overcome it?
how much data you train on? thanks

@vitrioil
Copy link
Owner

vitrioil commented Oct 5, 2020

Hi,

I did not solve the issue. I tried with around 20k audio clips for 2 person speech separation only. I would assume now that more data than that would be required. I did not experiment too much with the model, simply because the time it took for training 1 epoch was always in 1-2 day range, so couple of epochs would take weeks. This would change depending upon your GPU and VRAM availability. I would say more than 16GB would be helpful. So, a lot of opportunity there to tweak the model. I also found this. It could be helpful.

@vuthede
Copy link
Author

vuthede commented Oct 5, 2020

Hi, thanks for your quick reply.
In the paper, it seems like they do some preprocessing to remove noise in the input? Do u think it might help ?

@vitrioil
Copy link
Owner

vitrioil commented Oct 6, 2020

You mean to say while preparing the dataset? Well, I've seen someone mention that here. However, adding additional noise like AudioSet might help regularise.

@vuthede
Copy link
Author

vuthede commented Oct 7, 2020

Yeah, thanks,
Ah although you got the overfitting problem, I am curious that could your model can distinguish voice of 2 people in some extent?

@vitrioil
Copy link
Owner

vitrioil commented Oct 7, 2020

Yes, you could make out in certain instances who was the main speaker in the separated output. But, not always. Sometimes, it was only noise or mix of both the speakers. For the most part the output was noisy. All of this was also applicable to training data, but not to a great extent. As I said a lot of time is required for a model/dataset this big.

@JuanFMontesinos
Copy link

Probably related to #4

@MordehayM
Copy link

Hi,

I did not solve the issue. I tried with around 20k audio clips for 2 person speech separation only. I would assume now that more data than that would be required. I did not experiment too much with the model, simply because the time it took for training 1 epoch was always in 1-2 day range, so couple of epochs would take weeks. This would change depending upon your GPU and VRAM availability. I would say more than 16GB would be helpful. So, a lot of opportunity there to tweak the model. I also found this. It could be helpful.

Hi @vitrioil
Regarding the 20k audio clips, do you mean you downloaded 200 videos and extract from them the 20k audio clips? (200*199/2=19,000, this is the number of combination for creating the 20k audio clips given the 200 videos)

@vitrioil
Copy link
Owner

Hi @MordehayM ,

I believe it was 20k unique clips. 200C2 is indeed 19k, however not all combinations are considered.

There is a parameter: REMOVE_RANDOM_CHANCE (in audio_mixer_generator.py). This will prevent from combination from blowing up, otherwise there will be a lot of files created. By default the value is 0.9

Hence, I was not taking all combinations of files.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants