Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 7 additions & 0 deletions keras_nlp/utils/text_generation.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,8 @@
from absl import logging
from tensorflow import keras

from keras_nlp.api_export import keras_nlp_export


def _validate_prompt(prompt):
"""Helper function to validate input to text_generation utils."""
Expand Down Expand Up @@ -95,6 +97,7 @@ def _mask_tokens_after_end_token(
return tf.where(valid_indices, prompt, pad_token_id)


@keras_nlp_export("keras_nlp.utils.greedy_search")
def greedy_search(
token_probability_fn,
prompt,
Expand Down Expand Up @@ -210,6 +213,7 @@ def one_step(length, prompt):
return tf.squeeze(prompt) if input_is_1d else prompt


@keras_nlp_export("keras_nlp.utils.beam_search")
def beam_search(
token_probability_fn,
prompt,
Expand Down Expand Up @@ -400,6 +404,7 @@ def one_step(beams, beams_prob, length):
return tf.squeeze(prompt) if input_is_1d else prompt


@keras_nlp_export("keras_nlp.utils.random_search")
def random_search(
token_probability_fn,
prompt,
Expand Down Expand Up @@ -530,6 +535,7 @@ def one_step(length, prompt):
return tf.squeeze(prompt) if input_is_1d else prompt


@keras_nlp_export("keras_nlp.utils.top_k_search")
def top_k_search(
token_probability_fn,
prompt,
Expand Down Expand Up @@ -680,6 +686,7 @@ def one_step(length, prompt):
return tf.squeeze(prompt) if input_is_1d else prompt


@keras_nlp_export("keras_nlp.utils.top_p_search")
def top_p_search(
token_probability_fn,
prompt,
Expand Down