-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Prompt Image Alignment Experiment #6
Comments
This approach throw an error: which if check on their repo, it does not exists. I wonder how you work around that? |
I got the same "device-side assert triggered" |
For the "from llava.conversation import simple_conv_multimodal" error, you can simply use PROMPT = """You are LLaVA, a large language and vision assistant trained by UW Madison WAIV Lab.You are able to understand the visual content that the user provides, and assist the user with a variety of tasks using natural language.Follow the instructions carefully and explain your answers in detail.###Human: Hi!###Assistant: Hi there! How can I help you today?
###Human:""" instead of importing simple_conv_multimodal. |
Regarding “CUDA error: device-side assert triggered”, this happened when torch.embedding() tried to convert tokens into embeddings. Because LLaVA-server/llava_server/llava.py added special tokens to tokenizer, but it doesn't enlarge the embedding matrix of the model, which leads to this issue. We can resize the embedding matrix after adding special tokens. Add the following code in LLaVA-server/llava_server/llava.py line 38 may solve it. |
Hi Kevin, when I'm trying to reproduce the Prompt Alignment Experiment, I downloaded the llava_server codebase using weights from "liuhaotian/llava-v1.5-7b" first, when I run
I got KeyError: 'llava'. when loading weights
To handle this, I cloned the latest llava from https://github.com/haotian-liu/LLaVA and modified the llava_server/llava.py:
Now I'm testing this code on the machine with 3 A100 GPUs, it could load the weights and setup servers with app.py
However, when I use 2 GPUs for llava inference and run train.py on the other, I got:
"images = images.to("cuda", dtype=torch.float16)
RuntimeError: CUDA error: device-side assert triggered"
I also checked the nvidia smi that my processes were indeed on three GPUs separately. May I know if you have could help me with this? Thank you!
The text was updated successfully, but these errors were encountered: