Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loss curves #5

Open
Tomsen1410 opened this issue Apr 8, 2022 · 4 comments
Open

Loss curves #5

Tomsen1410 opened this issue Apr 8, 2022 · 4 comments
Labels
question Further information is requested

Comments

@Tomsen1410
Copy link

Hey, great work!
I just wanted to ask whether you might have the loss curves of your runs so that I can compare with my experiments a little bit?

@YuanGongND
Copy link
Owner

Hi Tom,

Thanks for your interest.

I didn't plot the loss curve, but I think I still have (pre)-training logs on my server. Which experiment are you looking for?

-Yuan

@YuanGongND YuanGongND added the question Further information is requested label Apr 8, 2022
@Tomsen1410
Copy link
Author

Tomsen1410 commented Apr 12, 2022

Thanks for the reply! I am looking for the pre-training run(s), i.e. the one trained on Librispeech + AudioSet.

@Tomsen1410
Copy link
Author

Bump!
Sorry to get back to the question. But I am currently trying to implement SSAST (well, its masked autoencoder counterpart, see https://arxiv.org/pdf/2203.16691.pdf) for music. For efficiency reasons, I tried to have the computed spectograms in FP16 format, but the reconstruction (generative) loss curve seems a bit weird. It first goes down really quickly, rises again to a given point and drops slowly again afterwards.

I just wanted to have some comparison in order to know what I should expect. Thanks in advance!

@YuanGongND
Copy link
Owner

YuanGongND commented Apr 26, 2022

Hi there,

I think this is our log (gen&dis objective, 400 masking patches, full AS + Librispeech). Unfortunately, I don't think we logged the generation loss but just the discrimination loss. The columns are defined at

result.append([train_acc_meter.avg, train_nce_meter.avg, acc_eval, nce_eval, optimizer.param_groups[0]['lr']])

For your question

reconstruction (generative) loss curve seems a bit weird. It first goes down really quickly, rises again to a given point and drops slowly again afterwards.

Could it be possible due to you add L_g and L_d together? Otherwise, the L_g on the training set should always drop.

-Yuan

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants