diff --git a/.travis.yml b/.travis.yml index cf43431d2fcc..d3f0487ae6ef 100644 --- a/.travis.yml +++ b/.travis.yml @@ -14,9 +14,9 @@ matrix: - python: 3.5 env: KERAS_BACKEND=tensorflow - python: 2.7 - env: KERAS_BACKEND=theano + env: KERAS_BACKEND=theano THEANO_FLAGS=optimizer=fast_compile - python: 3.5 - env: KERAS_BACKEND=theano + env: KERAS_BACKEND=theano THEANO_FLAGS=optimizer=fast_compile - python: 2.7 env: KERAS_BACKEND=cntk - python: 3.5 diff --git a/tests/keras/backend/backend_test.py b/tests/keras/backend/backend_test.py index f1da027094aa..2fc6867dcb91 100644 --- a/tests/keras/backend/backend_test.py +++ b/tests/keras/backend/backend_test.py @@ -839,7 +839,7 @@ def test_logsumexp(self, x_np, axis, keepdims, K): np.log(np.sum(np.exp(x_np), axis=axis, keepdims=keepdims)), rtol=1e-5) - @pytest.mark.parametrize('K', [KTH, KTF], ids=["KTH", "KTF"]) + @pytest.mark.parametrize('K', [KTF], ids=["KTF"]) def test_logsumexp_optim(self, K): ''' Check if optimization works. @@ -1387,12 +1387,6 @@ def test_batchnorm(self): ztf = KTF.eval(ztf) assert zth.shape == ztf.shape - # TODO remove this if statement when Theano without - # T.nnet.bn.batch_normalization_train is deprecated - zth2, _, _ = KTH._old_normalize_batch_in_training(xth, None, None, - reduction_axes=[0, 1, 2, 3]) - assert zth.shape == tuple(KTH.eval(zth2.shape)) - def test_ctc(self): # simplified version of TensorFlow's test