Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error when evaluating the qlora-merged model. #427

Open
vhientran opened this issue Nov 12, 2024 · 0 comments
Open

Error when evaluating the qlora-merged model. #427

vhientran opened this issue Nov 12, 2024 · 0 comments

Comments

@vhientran
Copy link

Hi Authors,

Thanks so much for releasing the great source code. I finetuned LlaMA-2-7B using QLoRA successfully on a node of 8GPUs A100 NVIDIA, and saved the merged model well. However, when loading this created model with vLLM, I got the error below:

model = vllm.LLM(
[rank0]:             ^^^^^^^^^
[rank0]:   File "/miniconda3/envs/open-instruct/lib/python3.11/site-packages/vllm/entrypoints/llm.py", line 177, in __init__
[rank0]:     self.llm_engine = LLMEngine.from_engine_args(
[rank0]:                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]:   File "/miniconda3/envs/open-instruct/lib/python3.11/site-packages/vllm/engine/llm_engine.py", line 573, in from_engine_args
[rank0]:     engine = cls(
[rank0]:              ^^^^
[rank0]: **KeyError: 'layers.11.mlp.down_proj.weight'**
Loading safetensors checkpoint shards:   0% Completed | 0/3 [00:00<?, ?it/s]

It seems the difference about the structure when loading the model. Could you please give me some suggestions to fix it? Many thanks for your help!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant