Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ListOps performance #15

Open
dido1998 opened this issue Jan 7, 2021 · 8 comments
Open

ListOps performance #15

dido1998 opened this issue Jan 7, 2021 · 8 comments

Comments

@dido1998
Copy link

dido1998 commented Jan 7, 2021

On running the ListOps task as-is from the repo, I got a validation performance similar to that reported in the paper but the test performance on results.json is very low:

{"accuracy": 0.17500001192092896, "loss": 3.032956123352051, "perplexity": 20.758506774902344}

I saw that the code is saving the model from the last checkpoint as compared to the model with the best validation performance. Could you detail the evaluation setup used in the paper i.e. for the paper do you evaluate the model from the last checkpoint of from the best validation checkpoint?

Thank you very much! :-)

@sihyun-yu
Copy link

Have you solved this problem? I have a similar issue.

@dido1998
Copy link
Author

Hi @sihyun-yu, I was not able to solve it

@apuaaChen
Copy link

I got a similar issue with the transformer_base. The evaluation accuracy curve is a little bit weird. The highest accuracy reaches 0.3359 at step 2.5k, then it drops to < 0.2. I directly use the default configurations.
image

@jinfengr
Copy link

It seems the issue is fixed with the latest code push. Please add comment if the issue still comes up.

@renebidart
Copy link

I found either lowering the learning rate or increasing the batch size was useful for this task. I think their hyperparameters are for a large effective batch size because they run on TPU.

@BalloutAI
Copy link

I am still getting the same problem, my validation during training is high on the listops, but when running test_only option, I am getting very low accuracy!

@BalloutAI
Copy link

The problem is that the data is shuffled every time the code is ran, so the tokens are changed when running the test script giving a random accuracy.

@yuzhenmao
Copy link

yuzhenmao commented Sep 15, 2022

The problem is that the data is shuffled every time the code is ran, so the tokens are changed when running the test script giving a random accuracy.

@BalloutAI Hi, I also found this issue: high training accuracy, low test accuracy; I also found if I run training process multiple times, sometimes the model cannot even converge. Could you explain your idea a little bit more? Thank you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants