You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The convert_llava_next_weights_to_hf.py script does not support converting the LLaVA-NeXT model based on Llama3-8B, llama3-llava-next-8b (announced here) meaning it is hard to load the weights of this model with LlavaNextForConditionalGeneration from the transformer library.
Motivation
Adding support would allow loading llama3-llava-next-8b, which is a strong multimodal model, with LlavaNextForConditionalGeneration included in the transformers lib, allowing full support for transformers functionality (in particular I'd like to use it with vLLM which only implements support for LlavaNextForConditionalGeneration rather than the model implementation from the LLaVA repo).
Your contribution
I have confirmed that modifying the script to add lmms-lab/llama3-llava-next-8b and setting
Feature request
The convert_llava_next_weights_to_hf.py script does not support converting the LLaVA-NeXT model based on Llama3-8B, llama3-llava-next-8b (announced here) meaning it is hard to load the weights of this model with
LlavaNextForConditionalGeneration
from the transformer library.Motivation
Adding support would allow loading
llama3-llava-next-8b
, which is a strong multimodal model, withLlavaNextForConditionalGeneration
included in the transformers lib, allowing full support for transformers functionality (in particular I'd like to use it with vLLM which only implements support forLlavaNextForConditionalGeneration
rather than the model implementation from the LLaVA repo).Your contribution
I have confirmed that modifying the script to add
lmms-lab/llama3-llava-next-8b
and settingworks and the model output seems sensible (although I'm unsure of the exact subtleties of the conversion and extra tokens). I have made a PR: #31395
The text was updated successfully, but these errors were encountered: