__init__(self, encoder_outputs, encoder_output_dim, encoder_lengths, vocab_size, attention_type, embedding_size, decoder_num_units, decoder_cells, residual_output_layers=None, name=None, weighted_encoder_outputs=None) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
apply(self, model, input_t, seq_lengths, states, timestep) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
apply_over_sequence(self, model, inputs, seq_lengths, initial_states) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
cell (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
decoder_output_dim (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
get_attention_weights(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
get_output_dim(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
get_outputs_with_grads(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
get_state_names(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
name (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
num_layers (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
output_indices (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
scope(self, name) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
use_attention (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |