Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Finetuning (Lora) #378

Closed
kevinintel opened this issue Jul 30, 2024 · 5 comments
Closed

Finetuning (Lora) #378

kevinintel opened this issue Jul 30, 2024 · 5 comments
Assignees
Labels
DEV features
Milestone

Comments

@kevinintel
Copy link
Collaborator

No description provided.

@endomorphosis
Copy link

can you be specific about what implementation you want to use for finetuning?

@nisten had finetuned https://huggingface.co/nisten/Biggie-SmoLlm-0.4B on the dual sapphire rapids system using https://github.com/axolotl-ai-cloud/axolotl?tab=readme-ov-file#train he got fp8 training working but i was unable to confirm with him whether the accelerate library was using AMX or AVX as the underlying instruction, and whether there are more performance optimizations that can be made.

The intention is that we will eventually need some sort of guardrail model, but he reports that right now the model training is "10x slower than a nvidia L4 24GB", and he will try to see what further optimizations he can make, and then write a readme for the purposes of OPEA.

@endomorphosis
Copy link

image

@nisten
Copy link

nisten commented Aug 1, 2024

I used code that would work with tim dettmers old bnb8bit linear optimizer but bits and bytes only supports that on GPU not cpu.
accelerate config supports mixed-precision float8_bfloat16 and it ran with it but i think it was just running bf16.
Training CPU only opens up someinteresting possibilities with using very very large batch sizes (like 256) in the training parameters.
Screenshot 2024-08-01 at 3 49 37 AM

@kevinintel
Copy link
Collaborator Author

we want to add fine tuning components and it can support LoRA.
But thanks for your sharing, we will create a basic fine tuning components first, you can raise a PR to add functionalities after it.
Thanks

lkk12014402 pushed a commit that referenced this issue Aug 8, 2024
* try with tgi-gaudi:2.0.0

Signed-off-by: chensuyue <[email protected]>

* try with text-generation-inference:2.1.0

Signed-off-by: chensuyue <[email protected]>

* add params for tgi-gaudi:2.0.0

Signed-off-by: chensuyue <[email protected]>

---------

Signed-off-by: chensuyue <[email protected]>
@XinyuYe-Intel
Copy link
Collaborator

Finetuning with LoRA is supported in this PR.

@kevinintel kevinintel added this to the v1.0 milestone Sep 3, 2024
@github-project-automation github-project-automation bot moved this to Done in OPEA Sep 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
DEV features
Projects
Status: Done
Development

No branches or pull requests

4 participants