From ac683746edeedc54e731fe2e7f2979dbbc472103 Mon Sep 17 00:00:00 2001 From: Guillaume Klein Date: Fri, 7 Aug 2020 16:09:35 +0200 Subject: [PATCH] Fix Layer class name in docstring --- tensorflow_addons/seq2seq/attention_wrapper.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tensorflow_addons/seq2seq/attention_wrapper.py b/tensorflow_addons/seq2seq/attention_wrapper.py index 349bbb7c30..457b0508e6 100644 --- a/tensorflow_addons/seq2seq/attention_wrapper.py +++ b/tensorflow_addons/seq2seq/attention_wrapper.py @@ -1664,8 +1664,8 @@ def __init__( `get_initial_state` which does not match the batch size of `initial_cell_state`, proper behavior is not guaranteed. name: Name to use when creating ops. - attention_layer: A list of `tf.tf.keras.layers.Layer` instances or a - single `tf.tf.keras.layers.Layer` instance taking the context + attention_layer: A list of `tf.keras.layers.Layer` instances or a + single `tf.keras.layers.Layer` instance taking the context and cell output as inputs to generate attention at each time step. If None (default), use the context as attention at each time step. If attention_mechanism is a list, attention_layer must be a list of