diff --git a/vllm/lora/models.py b/vllm/lora/models.py index a13b9346d17cc..5b7225bdc8f37 100644 --- a/vllm/lora/models.py +++ b/vllm/lora/models.py @@ -4,7 +4,7 @@ import os import re from dataclasses import dataclass, field -from typing import Any, Callable, Dict, List, Optional, Sequence, Type +from typing import Any, Callable, Dict, List, Optional, Sequence, Type, Union import safetensors.torch import torch @@ -219,7 +219,7 @@ def from_local_checkpoint( config["vllm_max_position_embeddings"] = max_position_embeddings peft_helper = PEFTHelper.from_dict(config) - unexpected_modules: List[str] + unexpected_modules: List[Union[list[str], str]] if os.path.isfile(lora_tensor_path): tensors: Dict[str, torch.Tensor] = {} # Find unexpected modules.