Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question for deployment #324

Open
iou2much opened this issue Nov 6, 2020 · 4 comments
Open

Question for deployment #324

iou2much opened this issue Nov 6, 2020 · 4 comments
Labels
enhancement New feature or request

Comments

@iou2much
Copy link

iou2much commented Nov 6, 2020

In my understanding, after export pb file and use CPP demo to transcribe. It doesn't use beam_search_decoder or wfst_decoder, it just output the transformer decoder result straightly. am I right?

If so, could anyone give some guidance for using beam_search or wfst in deployment mode? Thanks a lot

@Some-random
Copy link
Collaborator

The logit operation (addition/ assignment/ for loop) for beam_search_decoder and wfst_decoder are written with python. If you want to perform beam search with C++, there are two ways: 1 You need to create pbs that capture the network operations (encoder feature extraction, decoder step with encoder and previous states and inputs) and stitch them with C++ logits operations. 2 You can write the logit operation with tensorflow ops and freeze the whole graph to one pb. I believe the second option has already been implemented in MWER training of Speech Transformer.

@Some-random Some-random added the enhancement New feature or request label Nov 9, 2020
@iou2much
Copy link
Author

iou2much commented Nov 9, 2020

implemented in MWER training of Speech Transformer

Really? That's great. Let me check it out. Thank you

@iou2much
Copy link
Author

Hi, @Some-random and @hoyden . I've read the BatchBeamSearchLayer module in branch mwer.
Yet I still get some questions, could you help to interpret more here?
In BatchBeamSearchLayer, there's no scorer like CTCScorer, lm_model scorer. Do I need them in training stage or decoding stage? won't it help the performance?

@Some-random
Copy link
Collaborator

Hi, @Some-random and @hoyden . I've read the BatchBeamSearchLayer module in branch mwer.
Yet I still get some questions, could you help to interpret more here?
In BatchBeamSearchLayer, there's no scorer like CTCScorer, lm_model scorer. Do I need them in training stage or decoding stage? won't it help the performance?

BatchBeamSearchLayer is used in training stage, CTCScorer and lm_model scorer is not used in this stage. For decoding stage, adding these scorers will obviously boost the performance, but we haven't provided deployment with language model and CTC joint decoding yet

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants