Skip to content

Commit

Permalink
Merge pull request tensorflow#1425 from edouardfouche/patch-1
Browse files Browse the repository at this point in the history
Update lm_1b_eval.py (Python 3 compatibility)
  • Loading branch information
panyx0718 authored May 10, 2017
2 parents 2c44a4b + b2fc63b commit b312835
Showing 1 changed file with 5 additions and 4 deletions.
9 changes: 5 additions & 4 deletions lm_1b/lm_1b_eval.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,7 @@
"""
import os
import sys
import six

import numpy as np
import tensorflow as tf
Expand Down Expand Up @@ -83,7 +84,7 @@ def _LoadModel(gd_file, ckpt_file):
with tf.Graph().as_default():
sys.stderr.write('Recovering graph.\n')
with tf.gfile.FastGFile(gd_file, 'r') as f:
s = f.read()
s = f.read().decode()
gd = tf.GraphDef()
text_format.Merge(s, gd)

Expand Down Expand Up @@ -177,7 +178,7 @@ def _SampleModel(prefix_words, vocab):

prefix = [vocab.word_to_id(w) for w in prefix_words.split()]
prefix_char_ids = [vocab.word_to_char_ids(w) for w in prefix_words.split()]
for _ in xrange(FLAGS.num_samples):
for _ in six.moves.range(FLAGS.num_samples):
inputs = np.zeros([BATCH_SIZE, NUM_TIMESTEPS], np.int32)
char_ids_inputs = np.zeros(
[BATCH_SIZE, NUM_TIMESTEPS, vocab.max_word_length], np.int32)
Expand Down Expand Up @@ -230,7 +231,7 @@ def _DumpEmb(vocab):
sys.stderr.write('Finished softmax weights\n')

all_embs = np.zeros([vocab.size, 1024])
for i in range(vocab.size):
for i in six.moves.range(vocab.size):
input_dict = {t['inputs_in']: inputs,
t['targets_in']: targets,
t['target_weights_in']: weights}
Expand Down Expand Up @@ -269,7 +270,7 @@ def _DumpSentenceEmbedding(sentence, vocab):
inputs = np.zeros([BATCH_SIZE, NUM_TIMESTEPS], np.int32)
char_ids_inputs = np.zeros(
[BATCH_SIZE, NUM_TIMESTEPS, vocab.max_word_length], np.int32)
for i in xrange(len(word_ids)):
for i in six.moves.range(len(word_ids)):
inputs[0, 0] = word_ids[i]
char_ids_inputs[0, 0, :] = char_ids[i]

Expand Down

0 comments on commit b312835

Please sign in to comment.