Seq2Seq+attention代码
def attention(self, prev_state, enc_outputs):
#paramprev_state: the decoder hidden state at time i-1
#paramenc_outputs: the encoder outputs, a length 'T' list.
e_i= []
c_i= []
for output in enc_outputs:
atten_hidden= tf.tanh(tf.add(tf.matmul(prev_state, self.attention_W), tf.matmul(output, self.attention_U))
e_i_j= tf.matmul(atten_hidden, self.attention_V)
e_i.append(e_i_j)
e_i= tf.concat(e_i, axis=1)
alpha_i= tf.nn.softmax(e_i)
alpha_i= tf.split(alpha_i, self.num_steps, 1)
for alpha_i_j, output in zip(alpha_i, enc_outputs):
c_i_j= tf.multiply(alpha_i_j, output)
c_i.append(c_i_j)
c_i= tf.reshape(tf.concat(c_i, axis=1), [-1, self.num_steps, self.hidden_dim* 2])
c_i= tf.reduce_sum(c_i, 1)
return c_i
#对应的decode
def decode(self, cell, init_state, encoder_outputs, loop_function=None):
outputs = []
prev= None
state = init_state
for i, inpin enumerate(self.decoder_inputs_emb):#decoder_inputs_emb是tf.placeholder
#if loop_functionis not None and previs not None:
# with tf.variable_scope("loop_function", reuse=True):
# inp= loop_function(prev, i)
#if i> 0:
# tf.get_variable_scope().reuse_variables()
c_i= self.attention(state, encoder_outputs)
inp= tf.concat([inp, c_i], axis=1)
output, state = cell(inp, state)#原本没有attention的是decoder_input和state作为输入
outputs.append(output)
if loop_functionis not None:
prev= output
return outputs
#对应的decode
def decode(self, cell, init_state, encoder_outputs, loop_function=None):
outputs = []
prev= None
state = init_state
for i, inpin enumerate(self.decoder_inputs_emb):#decoder_inputs_emb是tf.placeholder
#if loop_functionis not None and previs not None:
# with tf.variable_scope("loop_function", reuse=True):
# inp= loop_function(prev, i)
#if i> 0:
# tf.get_variable_scope().reuse_variables()
c_i= self.attention(state, encoder_outputs)
inp= tf.concat([inp, c_i], axis=1)
output, state = cell(inp, state)#原本没有attention的是decoder_input和state作为输入
outputs.append(output)
if loop_functionis not None:
prev= output
return outputs