Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

❓ [QUESTION]Finetuning Validation Error Higher than Pre-training Error in Nequip #398

Open
ruoywang opened this issue Jan 1, 2024 · 0 comments
Labels
question Further information is requested

Comments

@ruoywang
Copy link

ruoywang commented Jan 1, 2024

Hi Nequip Team,

I am currently working with the Nequip framework for a project and have encountered an unexpected issue during the finetuning process. I would appreciate any insights or suggestions you have.

Issue Description:
I have been finetuning a model that was previously pre-trained. However, I noticed that the validation error during finetuning is consistently higher than the error observed during the pre-training phase. To investigate further, I set the validation set as the finetuning training set. Surprisingly, even under these conditions, the validation error (especially in the earliest epochs) of new training remains higher than it was in the pre-training.

Intuitively, when the training set is the same as the validation set, the validation error should be lower than the pre-training error.

I would like to understand whether this higher validation error observed during finetuning is a common result inherent to the algorithm itself or if it might be due to some improper settings on my part. If it's the latter, I am willing to provide my input and output files for further analysis.

Thank you very much for your time and assistance.

Best regards,
Ruoyu

@ruoywang ruoywang added the question Further information is requested label Jan 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

1 participant