From 04df1d52d77f41597c2d5586e04d083d652b7f05 Mon Sep 17 00:00:00 2001 From: Jee Jee Li Date: Fri, 12 Sep 2025 16:37:19 +0000 Subject: [PATCH 1/2] Done Signed-off-by: Jee Jee Li --- vllm/v1/worker/lora_model_runner_mixin.py | 20 ++++++++++---------- 1 file changed, 10 insertions(+), 10 deletions(-) diff --git a/vllm/v1/worker/lora_model_runner_mixin.py b/vllm/v1/worker/lora_model_runner_mixin.py index f2ebd5e10210..34f1bbba1c14 100644 --- a/vllm/v1/worker/lora_model_runner_mixin.py +++ b/vllm/v1/worker/lora_model_runner_mixin.py @@ -63,8 +63,7 @@ def load_lora_model(self, model: nn.Module, model_config: ModelConfig, def _set_active_loras(self, prompt_lora_mapping: tuple[int, ...], token_lora_mapping: tuple[int, ...], lora_requests: set[LoRARequest]) -> None: - if not self.lora_manager: - raise RuntimeError("LoRA is not enabled.") + self._ensure_lora_enabled() # Set is_prefill to True, so we always use the SGMV kernels on # non-cuda platforms. @@ -75,6 +74,11 @@ def _set_active_loras(self, prompt_lora_mapping: tuple[int, ...], is_prefill=True) self.lora_manager.set_active_adapters(lora_requests, lora_mapping) + def _ensure_lora_enabled(self) -> None: + if not hasattr(self, "lora_manager"): + raise RuntimeError( + "LoRA is not enabled. Use --enable-lora to enable LoRA.") + def set_active_loras(self, input_batch: InputBatch, num_scheduled_tokens: np.ndarray) -> None: @@ -172,21 +176,17 @@ def maybe_remove_all_loras(self, lora_config: Optional[LoRAConfig]): self.lora_manager.remove_all_adapters() def add_lora(self, lora_request: LoRARequest) -> bool: - if not self.lora_manager: - raise RuntimeError("LoRA is not enabled.") + self._ensure_lora_enabled() return self.lora_manager.add_adapter(lora_request) def remove_lora(self, lora_id: int) -> bool: - if not self.lora_manager: - raise RuntimeError("LoRA is not enabled.") + self._ensure_lora_enabled() return self.lora_manager.remove_adapter(lora_id) def pin_lora(self, lora_id: int) -> bool: - if not self.lora_manager: - raise RuntimeError("LoRA is not enabled.") + self._ensure_lora_enabled() return self.lora_manager.pin_adapter(lora_id) def list_loras(self) -> set[int]: - if not self.lora_manager: - raise RuntimeError("LoRA is not enabled.") + return self.lora_manager.list_adapters() From 15d6661b57cfa3074c627102b3e80697d70576b0 Mon Sep 17 00:00:00 2001 From: Jee Jee Li Date: Fri, 12 Sep 2025 16:52:31 +0000 Subject: [PATCH 2/2] Fix Signed-off-by: Jee Jee Li --- vllm/v1/worker/lora_model_runner_mixin.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/v1/worker/lora_model_runner_mixin.py b/vllm/v1/worker/lora_model_runner_mixin.py index 34f1bbba1c14..01d5f0525c4e 100644 --- a/vllm/v1/worker/lora_model_runner_mixin.py +++ b/vllm/v1/worker/lora_model_runner_mixin.py @@ -188,5 +188,5 @@ def pin_lora(self, lora_id: int) -> bool: return self.lora_manager.pin_adapter(lora_id) def list_loras(self) -> set[int]: - + self._ensure_lora_enabled() return self.lora_manager.list_adapters()