attention-model Questions
1
Solved
I'm trying to implement torch.nn.TransformerEncoder with a src_key_padding_mask not equal to none. Imagine the input is of the shape src = [20, 95] and the binary padding mask has the shape src_mas...
Terresaterrestrial asked 16/6, 2020 at 0:43
2
Solved
I've tried to build a sequence to sequence model to predict a sensor signal over time based on its first few inputs (see figure below)
The model works OK, but I want to 'spice things up' and try ...
Irreconcilable asked 12/5, 2020 at 16:56
1
Based on this article, I wrote this model:
enc_in=Input(shape=(None,in_alphabet_len))
lstm=LSTM(lstm_dim,return_sequences=True,return_state=True,use_bias=False)
enc_out,h,c=lstm(enc_in)
dec_in=Inp...
Inanimate asked 8/11, 2017 at 9:25
1
Solved
I'm using (keras-self-attention) to implement attention LSTM in KERAS. How can I visualize the attention part after training the model? This is a time series forecasting case.
from keras.models im...
Nonjoinder asked 12/10, 2019 at 17:47
0
I would like to implement attention to a trained image classification CNN model. For example, there are 30 classes and with the Keras CNN, I obtain for each image the predicted class. However, to v...
Colon asked 16/7, 2019 at 14:15
2
Solved
To be clear, I am referring to "self-attention" of the type described in Hierarchical Attention Networks for Document Classification and implemented many places, for example: here. I am not referri...
Yentai asked 27/3, 2018 at 21:27
1
I want to provide a mask, the same size as the input image and adjust the weights learned from the image according to this mask (similar to attention, but pre-computed for each image input). How ca...
Gerrald asked 27/2, 2019 at 13:43
1
Solved
I am trying to implement the attention described in Luong et al. 2015 in PyTorch myself, but I couldn't get it work. Below is my code, I am only interested in the "general" attention case for now. ...
Liking asked 28/5, 2018 at 18:41
2
Solved
The attention mechanism of LSTM is a straight softmax feed forward network that takes in the hidden states of each time step of the encoder and the decoder's current state.
These 2 steps seems to ...
Coplanar asked 8/6, 2017 at 18:48
0
How exactly is tf.contrib.rnn.AttentionCellWrapper used? Can someone give a piece of example code?
Specifically, I only managed to make the following
fwd_cell = tf.contrib.rnn.AttentionCellWrap...
Caryncaryo asked 25/5, 2017 at 1:41
1
Solved
Is there a way to visualize the attention weights on some input like the figure in the link above(from Bahdanau et al., 2014), in TensorFlow's seq2seq models? I have found TensorFlow's github iss...
Concerto asked 15/11, 2016 at 3:34
0
While attempting to replicate the section 3.1 in Incorporating Discrete Translation Lexicons into Neural MT in paddle-paddle
I tried to have a static matrix that I'll need to load into the seqTose...
Folberth asked 17/10, 2016 at 5:57
© 2022 - 2024 — McMap. All rights reserved.