Proper way use MULTI GPUs with DEEPSPEED when each GPU does not have enough VRAM? #1369
Unanswered
atipasvanund
asked this question in
Q&A
Replies: 1 comment
-
You should be able to using zero3 with offload. Alternatively , make sure to adjust optim etc. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
What's the proper way to fine-tune on MULTI GPUs with DEEPSPEED when each GPU does not have enough VRAM? For example If I wish to use 4 x RTX 4090 to fine-tune a 7B model without LORA or Quantization. Is this possible? I tried zero1.json all the way to zero3_bf16.json and get various error, including out of CUDA memory. Can someone please help me?
Beta Was this translation helpful? Give feedback.
All reactions