-
Notifications
You must be signed in to change notification settings - Fork 143
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Finetuning (Lora) #378
Comments
can you be specific about what implementation you want to use for finetuning? @nisten had finetuned https://huggingface.co/nisten/Biggie-SmoLlm-0.4B on the dual sapphire rapids system using https://github.com/axolotl-ai-cloud/axolotl?tab=readme-ov-file#train he got fp8 training working but i was unable to confirm with him whether the accelerate library was using AMX or AVX as the underlying instruction, and whether there are more performance optimizations that can be made. The intention is that we will eventually need some sort of guardrail model, but he reports that right now the model training is "10x slower than a nvidia L4 24GB", and he will try to see what further optimizations he can make, and then write a readme for the purposes of OPEA. |
we want to add fine tuning components and it can support LoRA. |
* try with tgi-gaudi:2.0.0 Signed-off-by: chensuyue <[email protected]> * try with text-generation-inference:2.1.0 Signed-off-by: chensuyue <[email protected]> * add params for tgi-gaudi:2.0.0 Signed-off-by: chensuyue <[email protected]> --------- Signed-off-by: chensuyue <[email protected]>
Finetuning with LoRA is supported in this PR. |
No description provided.
The text was updated successfully, but these errors were encountered: