-
Notifications
You must be signed in to change notification settings - Fork 377
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TopKDecoder #177
Comments
hi, I am studying the code and have similar doubts. However, can you be clear what you mean by decoder_output? do you actually mean log_softmax_output? |
@JojoFisherman Yeah, I mean the output probability of decoder, i.e. log_softmax_output. |
I have the same question. It surprised me that no one has answered this. If theres really something wrong in the beam search, surely it will output some weird sequence. Do you have any conclusion about this? |
It seems some issues have referred that beam search doesn't work correctly. Unfortunately, maybe this repo is not active maintained now. Currently, I use fairseq (pytorch version) to conduct some related experiments. |
I studied the codes these days, and I thought you can use the torch.repeat_interleave. Such as follow: |
I had the problem with batch_size > 1, but after applying this comment, then it works now. Thank you!! |
Hi,
I wonder if rnn.forward_step changes the order of (batch_size*self.k) dimension ?
With the code about initializing sequence_scores:
and in each step:
It seems like sequence_scores is updated as (assume that selk.k = 3):
If hidden and inflated_encoder_outputs should be calculated as follow?
The text was updated successfully, but these errors were encountered: