Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Attention-based models do not work #16

Open
hirofumi0810 opened this issue Oct 12, 2017 · 3 comments
Open

Attention-based models do not work #16

hirofumi0810 opened this issue Oct 12, 2017 · 3 comments

Comments

@hirofumi0810
Copy link
Owner

I'm refactoring attention-based models, so they do not work now!
Please wait for a few days.

@wolverineq
Copy link

Good to hear! and hope a multi-gpu training version will be added since training of attention-based model is quite time-consuming...

@hirofumi0810
Copy link
Owner Author

now finished (greedy decoding)

TODO:

  • beam search
  • multi GPUs version

@hirofumi0810 hirofumi0810 changed the title Attention-based models do now work Attention-based models do not work Oct 27, 2017
@CyanHillFox
Copy link

@hirofumi0810 @hlthu
Does it work now? It seems that the preprocessed data created by this repo doesn't meet the requirement.

I tried the timit experiment with attention-based model, but it seems that the structure of input data doesn't match. For example, in timit/data/load_dataset_attention.py, it expects that label and input data are in "label" and "input" folder, but in the preprocessed data, they are in "dataset" and "feature". And that the labels are in .csv file rather than .npy file.
Maybe there should be an update?

Sincerely.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants