| __init__(self, encoder_outputs, encoder_output_dim, encoder_lengths, vocab_size, attention_type, embedding_size, decoder_num_units, decoder_cells, residual_output_layers=None, name=None, weighted_encoder_outputs=None) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| apply(self, model, input_t, seq_lengths, states, timestep) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| apply_over_sequence(self, model, inputs, seq_lengths, initial_states) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| cell (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| decoder_output_dim (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| get_attention_weights(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| get_output_dim(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| get_outputs_with_grads(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| get_state_names(self) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| name (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| num_layers (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| output_indices (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| scope(self, name) (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |
| use_attention (defined in caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder) | caffe2.python.models.seq2seq.seq2seq_util.LSTMWithAttentionDecoder | |