diff --git a/extension/android/src/main/java/org/pytorch/executorch/LlamaModule.java b/extension/android/src/main/java/org/pytorch/executorch/LlamaModule.java index 2fe826fc0af..6de26bc7fe8 100644 --- a/extension/android/src/main/java/org/pytorch/executorch/LlamaModule.java +++ b/extension/android/src/main/java/org/pytorch/executorch/LlamaModule.java @@ -173,7 +173,7 @@ public long prefillPrompt(String prompt, long startPos, int bos, int eos) { * @param prompt The text prompt to LLaVA. * @param seqLen The total sequence length, including the prompt tokens and new tokens. * @param startPos The starting position in KV cache of the input in the LLM. - * @param llamaCallback callback object to receive results. + * @param callback callback object to receive results. * @param echo indicate whether to echo the input prompt or not. * @return The error code. */