Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Memory leak in shared GPU memory #107

Open
viper7882 opened this issue Oct 5, 2024 · 1 comment
Open

Memory leak in shared GPU memory #107

viper7882 opened this issue Oct 5, 2024 · 1 comment

Comments

@viper7882
Copy link

Hi admin,

After running Lag-Llama Demo, the dedicated GPU memory is freed up but the shared GPU memory is stuck at fully occupied. I suspect it has not been freed up properly.

One point to note is during the early execution of the demo, I saw the shared memory was fluctuating from zero to max and vice versa. Hence, I strongly believe shared memory is fully controllable by Lag-Llama.

Screenshot of demo post execution using Jupyter Notebook:
Lag-llama did not free up shared GPU memory

You might want to check it out.

@lsf123-ai
Copy link

Hello, I'd like to ask you for some advice. My computer is configured with CUDA and the corresponding PyTorch. When running script files, GPU available: True (cuda), used: True, but in fact, when I check my computer's performance, it's not using the GPU. Could you please tell me what the reason might be? I'm directly running a pre-trained script file without specifically modifying any code related to the GPU. If I need to modify the code, which part should I change? Your guidance would be greatly appreciated.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants