Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MusicCap Dataset Testing: Overfitting Issues in Model Predictions #9

Open
oyzh888 opened this issue Jan 31, 2024 · 3 comments
Open

Comments

@oyzh888
Copy link

oyzh888 commented Jan 31, 2024

Hello,

I appreciate your excellent work and have a question regarding the testing process, specifically on how to ensure proper testing without falling into the trap of overfitting.

We conducted a test using the MusicCap dataset (https://huggingface.co/datasets/google/MusicCaps), which contains approximately 5.52K samples, somewhat akin to the 60K mentioned in your paper.

However, we encountered an issue where some samples appear to be "overfitting". Is this a normal occurrence? For instance, we observed cases where the model's prediction exactly matches the label in the MusicCap dataset.

One example involves the YouTube video with the ID -FFx68qSAuY (https://www.youtube.com/watch?v=-FFx68qSAuY).
Audio file(you should uncompress it):
-FFx68qSAuY.wav.zip

The model predicted:

{
  'text': 'This is a punk rock music piece. There are male vocals singing in a grunt-like manner. The melody is being played by an electric guitar while a bass guitar plays in the background. The rhythm consists of a slightly fast-paced rock acoustic drum beat. The piece has an aggressive atmosphere. It could be used in the soundtrack of an action-filled video game.',
  'time': '0:00-10:00'
}

For our tests, we used the following code: https://github.com/seungheondoh/lp-music-caps/blob/main/lpmc/music_captioning/captioning.py#L52, and executed the command:

python3 captioning.py --audio_path ../music_cap/lp-music-caps/lpmc/music_captioning/workspace/audio/-FFx68qSAuY.wav

Additionally, we noticed similar overfitting issues in approximately 10% of the samples, including these YouTube links:
https://www.youtube.com/watch?v=PpJKo-JPVU0
https://www.youtube.com/watch?v=p0oRrGDrQQw
Could you provide insights or guidance on this matter?

Thank you.

@diggerdu
Copy link

diggerdu commented Mar 4, 2024

@seungheondoh
Hello,
I really appreciate your excellent work. But unknown train-test splits make it hard to follow up.

@seungheondoh
Copy link
Owner

seungheondoh commented Apr 29, 2024

@oyzh888
Sorry for the late reply. My model was trained with MusicCaps Training Split.
Therefore, if you infer with MusicCaps, you may get similar results.
The uploaded data is MusicCaps Test Split?

@seungheondoh
Copy link
Owner

@diggerdu please check #4

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants