Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

运行 GLM-10B 的最低配置是多少? #191

Open
nguyenvanhoangphuc opened this issue Sep 8, 2023 · 1 comment
Open

运行 GLM-10B 的最低配置是多少? #191

nguyenvanhoangphuc opened this issue Sep 8, 2023 · 1 comment

Comments

@nguyenvanhoangphuc
Copy link

我正在使用一台16GB内存的CPU和一台4GB内存的GPU,但由于内存溢出而无法运行GLM-10B。 请问我需要什么确切的配置来运行它,而不会遇到内存溢出错误?

@fluxlinkage
Copy link

glm-10b-chinese模型:
直接GPU推理需要大约20GB显存。
使用如下方法可以8位量化,显存需要十几GB(具体多少忘了,11GB到15GB之间吧)。
直接CPU推理需要大约50-60GB内存。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants