From fccbb3cf96402185606ba3df920d06be527783c1 Mon Sep 17 00:00:00 2001 From: sayakpaul Date: Wed, 3 Jul 2024 09:02:45 +0530 Subject: [PATCH] add dummy lora attention processors to prevent failures in other libs --- src/diffusers/models/attention_processor.py | 20 ++++++++++++++++++++ 1 file changed, 20 insertions(+) diff --git a/src/diffusers/models/attention_processor.py b/src/diffusers/models/attention_processor.py index 8784dcda4b6e..9d495695e330 100644 --- a/src/diffusers/models/attention_processor.py +++ b/src/diffusers/models/attention_processor.py @@ -2775,6 +2775,26 @@ def __call__( return hidden_states +class LoRAAttnProcessor: + def __init__(self): + pass + + +class LoRAAttnProcessor2_0: + def __init__(self): + pass + + +class LoRAXFormersAttnProcessor: + def __init__(self): + pass + + +class LoRAAttnAddedKVProcessor: + def __init__(self): + pass + + ADDED_KV_ATTENTION_PROCESSORS = ( AttnAddedKVProcessor, SlicedAttnAddedKVProcessor,