Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Confused due to epochs number #2

Open
powerswing opened this issue Sep 6, 2017 · 0 comments
Open

Confused due to epochs number #2

powerswing opened this issue Sep 6, 2017 · 0 comments

Comments

@powerswing
Copy link

powerswing commented Sep 6, 2017

Hello.

First of all, I am really sorry for disturbing you, but nonetheless I would like to make some inquiries concerning the dataset, the model and the training process:

  • To me, one million epochs sounds incredible and confusing. The fact is I have seen different examples of reproducing texts made by different authors with Keras. There, the usual scope of epochs is between 20 and 60 (for instance, here it takes 20 epochs to make the model produce a somewhat coherent text). I have seen your code and it seems to me that the epoch is considered done is the same as using Keras, i. e. the next epoch starts after all the batches have been passed to .fit(). I suppose your corpus is around ~1 mil. characters, and passing them by 200-sized batches for a million epochs would definitely take more than 2 hours on your GPU (Or is it really 2 hours? I personally do not have the same gear. I use 8 CPU on google sdk). So, the question is when is the batch considered completed? first priority question

  • What are the shapes of X and y variables? second priority question

  • What is the corpus (content) length of the dataset? In other words, how many characters does it contain in total? third priority question (may leave it be if the second question is answered)

Thank you in advance.
E.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant