From cac9d9a8afb0d26bae87fa0ab71eef563fd6dbb4 Mon Sep 17 00:00:00 2001 From: Wagner Bruna Date: Thu, 13 Nov 2025 13:29:58 -0300 Subject: [PATCH] fix: typo in --lora-apply-mode help --- examples/cli/README.md | 4 ++-- examples/cli/main.cpp | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/examples/cli/README.md b/examples/cli/README.md index 84df1a1e..cc5896fd 100644 --- a/examples/cli/README.md +++ b/examples/cli/README.md @@ -103,7 +103,7 @@ Options: contain any quantized parameters, the at_runtime mode will be used; otherwise, immediately will be used.The immediately mode may have precision and compatibility issues with quantized parameters, but it usually offers faster inference - speed and, in some cases, lower memory usageThe at_runtime mode, on the other + speed and, in some cases, lower memory usage. The at_runtime mode, on the other hand, is exactly the opposite. --scheduler denoiser sigma scheduler, one of [discrete, karras, exponential, ays, gits, smoothstep, sgm_uniform, simple], default: discrete @@ -119,4 +119,4 @@ Options: --vae-relative-tile-size relative tile size for vae tiling, format [X]x[Y], in fraction of image size if < 1, in number of tiles per dim if >=1 (overrides --vae-tile-size) --preview preview method. must be one of the following [none, proj, tae, vae] (default is none) -``` \ No newline at end of file +``` diff --git a/examples/cli/main.cpp b/examples/cli/main.cpp index a2df0949..f7c6488d 100644 --- a/examples/cli/main.cpp +++ b/examples/cli/main.cpp @@ -1144,7 +1144,7 @@ void parse_args(int argc, const char** argv, SDParams& params) { "the way to apply LoRA, one of [auto, immediately, at_runtime], default is auto. " "In auto mode, if the model weights contain any quantized parameters, the at_runtime mode will be used; otherwise, immediately will be used." "The immediately mode may have precision and compatibility issues with quantized parameters, " - "but it usually offers faster inference speed and, in some cases, lower memory usage" + "but it usually offers faster inference speed and, in some cases, lower memory usage. " "The at_runtime mode, on the other hand, is exactly the opposite.", on_lora_apply_mode_arg}, {"",