Skip to content

Commit

Permalink
Small cleanup
Browse files Browse the repository at this point in the history
  • Loading branch information
wouterkool committed Jun 25, 2018
1 parent 86f486e commit 9f05ec8
Showing 1 changed file with 0 additions and 4 deletions.
4 changes: 0 additions & 4 deletions attention_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -144,14 +144,10 @@ def _inner(self, input, embeddings):
# Compute keys, values for the glimpse and keys for the logits once as they can be reused in every step
fixed_context, attention_node_data_fixed = self._precompute(embeddings)

context_list = []

# Perform decoding steps
i = 0
while not self._is_finished(i, state):

context_list.append(state)

log_p, mask = self._get_log_p(embeddings, fixed_context, attention_node_data_fixed, state)

# Select the indices of the next nodes in the sequences, result (batch_size) long
Expand Down

0 comments on commit 9f05ec8

Please sign in to comment.