From c24fd8780dcc06fd5456ca9e9335b669ccb8b5be Mon Sep 17 00:00:00 2001 From: Guillaume Klein Date: Wed, 12 Feb 2020 18:02:06 +0100 Subject: [PATCH] Re-enable Luong monotonic tests --- tensorflow_addons/seq2seq/attention_wrapper_test.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/tensorflow_addons/seq2seq/attention_wrapper_test.py b/tensorflow_addons/seq2seq/attention_wrapper_test.py index 6980d90d54..95cb567ece 100644 --- a/tensorflow_addons/seq2seq/attention_wrapper_test.py +++ b/tensorflow_addons/seq2seq/attention_wrapper_test.py @@ -826,8 +826,6 @@ def testBahdanauMonotonicNormalized(self): create_attention_kwargs=create_attention_kwargs) def testLuongMonotonicNotNormalized(self): - self.skipTest( - "Resolve https://github.com/tensorflow/addons/issues/781") create_attention_mechanism = wrapper.LuongMonotonicAttention expected_final_output = basic_decoder.BasicDecoderOutput( @@ -862,8 +860,6 @@ def testLuongMonotonicNotNormalized(self): expected_final_alignment_history=expected_final_alignment_history) def testLuongMonotonicScaled(self): - self.skipTest( - "Resolve https://github.com/tensorflow/addons/issues/781") create_attention_mechanism = wrapper.LuongMonotonicAttention create_attention_kwargs = {"scale": True}