diff --git a/swift/llm/model/model/deepseek.py b/swift/llm/model/model/deepseek.py index 3fb4356c0f..f711c86261 100644 --- a/swift/llm/model/model/deepseek.py +++ b/swift/llm/model/model/deepseek.py @@ -195,7 +195,7 @@ def get_model_tokenizer_deepseek_vl2(model_dir: str, *args, **kwargs): if not local_repo_path: local_repo_path = git_clone_github('https://github.com/deepseek-ai/DeepSeek-VL2') sys.path.append(os.path.join(local_repo_path)) - from deepseek_vl.models import DeepseekVLV2Processor, DeepseekVLV2ForCausalLM + from deepseek_vl2.models import DeepseekVLV2Processor, DeepseekVLV2ForCausalLM processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_dir) return _get_deepseek_vl(processor, 'language', model_dir, *args, **kwargs) diff --git a/swift/llm/template/template/deepseek.py b/swift/llm/template/template/deepseek.py index a37a9da2df..7c5b84d4e2 100644 --- a/swift/llm/template/template/deepseek.py +++ b/swift/llm/template/template/deepseek.py @@ -126,7 +126,7 @@ class DeepseekVL2Template(DeepseekVLTemplate): image_placeholder = ['\n'] def _encode(self, inputs: StdTemplateInputs) -> Dict[str, Any]: - from deepseek_vl.models.processing_deepseek_vl_v2 import VLChatProcessorOutput + from deepseek_vl2.models.processing_deepseek_vl_v2 import VLChatProcessorOutput encoded = Template._encode(self, inputs) images = inputs.images processor = self.processor