-
Notifications
You must be signed in to change notification settings - Fork 117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Clarify requirements #12
Comments
Hi @vid! Totally 128 Gb of RAM, I assume, is ok. Moreover, you can not install more than 128 Gb in a typical desktop, even i9-13900k supports only 128 Gb max. The systems that allow more RAM, immediately costs twice or more. CUDA card is not so important for this repo, it just runs LLaMA layer by layer, so may be even 1080ti could handle this. If you have 3090ti, may be it is better to find another repos that would act faster with your cool card and doesn't require so much RAM. |
Thank you for that response! |
What about 64gig ecc server ram |
If you really need ECC RAM you could buy a server or workstation class system, maybe used. DDR RAM has on-chip error checking. Normally ECC isn't that important, worst case you can run a program twice to see if you get the same result, though it's even more questionable with DL. |
What would be the amount of gigs of vram to run this model on a GPU? I'm considering buying an A100 with the company's resources 😈 |
@tallesairan 1 Gb on a GeForce 710 should be enough , just it will be a little bit slowly. This repo feeds layers one by one and the greatest layer is tokenizer which is 500 Mb :) |
Is it possible to trade off the lack of RAM in favor of GPU RAM ? |
@nopium HF version allows GPU offloading, but we still need a lot of RAM. |
Hi, I am ordering some RAM to work with LLAMA when I take a break in a few weeks. The README for this repo says "64 or better 128 Gb of RAM (192 or 256 would be perfect)". Is this alongside a CUDA card? I have a 3090. I can order up to 192GB of RAM, if it makes a big difference. Will it?
Thanks!
The text was updated successfully, but these errors were encountered: