Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Were both training and test splits from original datasets (e.g., AudioSet, VggSound) used to train CLAP? #97

Open
ttgeng233 opened this issue Aug 10, 2023 · 1 comment

Comments

@ttgeng233
Copy link

I want to know the exact splits of AudioSet or VggSound used to train the CLAP. Because many audio-related datasets for downstream tasks were collected from these two large-scale datasets, if all their test data were seen during the pre-training stage, the evaluation results would be unconvincing.

@YuchenHui22314
Copy link
Collaborator

While evaluating, we manually eliminate those examples already seen in the pretraining stage. For example, while testing on ESC-50, we eliminated all overlaps with freesound and audioset.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants