-
-
Notifications
You must be signed in to change notification settings - Fork 897
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Peft deepspeed resume #1227
Peft deepspeed resume #1227
Conversation
Looks like it does the same as huggingface/transformers#28746 |
Thanks for that. I'll wait for that to get merged 🤞. Hard to keep track.of everything upstream. |
Once this is fixed upstream, we can remove the monkeypatch from this PR, but I think we still need to handle the lora_model_dir part. |
@manishiitg this was fixed upstream, can you confirm if the upstream fix works for you? |
5594554
to
839637c
Compare
looks like we need to handle some changes from this PR too huggingface/transformers#26610 |
* import deepspeed integration * monkeypatch peft adapater with deepspeed for resume from checkpoint * fix patch * fix patches attempt 2 * make sure to set lora_model_dir * skip pylint for deepspeed.utils * pick up upstream fix in transformers * remove monkeypatch for deepspeed/peft fix * no need to set the lora_model_dir on resume * unset load_in_*bit when using quant config * guard before del * better handling of load_in* kwargs
Fixes #1134