From fcaa98009b739d7bee7552464839060369a8a22e Mon Sep 17 00:00:00 2001 From: mobicham Date: Mon, 15 Jan 2024 12:17:12 +0000 Subject: [PATCH] fix lora load filename --- hqq/core/peft.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/hqq/core/peft.py b/hqq/core/peft.py index 2292eba..e17dd8c 100755 --- a/hqq/core/peft.py +++ b/hqq/core/peft.py @@ -332,7 +332,7 @@ def load_lora_weights(cls, model, filename, base_class=None, verbose=True): #Base classs base_class = cls.get_base_class(model, base_class) - lora_global_params = torch.load(file, map_location='cpu') + lora_global_params = torch.load(filename, map_location='cpu') def _patch_linear_load_weights(layer, patch_params, return_layer=True): if(is_hqq_lora_layer(layer)):