Doesn't work. The APP would stuck on the LLM loading page, and crash
private val mediaPipeLanguageModelOptions: LlmInferenceOptions =
LlmInferenceOptions.builder().setModelPath(
baseFolder + GEMMA_MODEL
).setPreferredBackend(LlmInference.Backend.GPU).setMaxTokens(4096).build()
Though on CPU it works normally
private val mediaPipeLanguageModelOptions: LlmInferenceOptions =
LlmInferenceOptions.builder().setModelPath(
baseFolder + GEMMA_MODEL
).setPreferredBackend(LlmInference.Backend.CPU).setMaxTokens(4096).build()
Doesn't work. The APP would stuck on the LLM loading page, and crash
Though on CPU it works normally