Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLaVA support #478

Closed
lifo9 opened this issue Feb 12, 2024 · 7 comments
Closed

LLaVA support #478

lifo9 opened this issue Feb 12, 2024 · 7 comments

Comments

@lifo9
Copy link

lifo9 commented Feb 12, 2024

Feature request

The support is already present in huggingface/transformers.

But when I try to export LLaVA model to neuron format, it throws the following error:

optimum-cli export neuron --model liuhaotian/llava-v1.6-vicuna-7b --disable-validation /llava/
File "/usr/local/lib/python3.10/dist-packages/optimum/exporters/tasks.py", line 1140, in get_supported_tasks_for_model_type
    raise KeyError(
KeyError: "llava is not supported yet for transformers. Only ['audio-spectrogram-transformer', 'albert', 'bart', 'beit', 'bert', 'blenderbot', 'blenderbot-small', 'bloom', 'camembert', 'clip', 'clip-text-model', 'clip-text-with-projection', 'codegen', 'convbert', 'convnext', 'convnextv2', 'cvt', 'data2vec-text', 'data2vec-vision', 'data2vec-audio', 'deberta', 'deberta-v2', 'deit', 'detr', 'distilbert', 'donut', 'donut-swin', 'dpt', 'electra', 'encoder-decoder', 'esm', 'falcon', 'flaubert', 'glpn', 'gpt2', 'gpt-bigcode', 'gptj', 'gpt-neo', 'gpt-neox', 'groupvit', 'hubert', 'ibert', 'imagegpt', 'layoutlm', 'layoutlmv3', 'lilt', 'levit', 'longt5', 'marian', 'mbart', 'mistral', 'mobilebert', 'mobilevit', 'mobilenet-v1', 'mobilenet-v2', 'mpnet', 'mpt', 'mt5', 'm2m-100', 'nystromformer', 'owlvit', 'opt', 'llama', 'pegasus', 'perceiver', 'phi', 'pix2struct', 'poolformer', 'regnet', 'resnet', 'default-timm-config', 'roberta', 'roformer', 'sam', 'segformer', 'sentence-transformers-clip', 'sentence-transformers-transformer', 'sew', 'sew-d', 'speech-to-text', 'speecht5', 'splinter', 'squeezebert', 'swin', 'swin2sr', 't5', 'trocr', 'unet', 'unispeech', 'unispeech-sat', 'vae-encoder', 'vae-decoder', 'vision-encoder-decoder', 'vit', 'wavlm', 'wav2vec2', 'wav2vec2-conformer', 'whisper', 'xlm', 'xlm-roberta', 'yolos'] are supported. If you want to support llava please propose a PR or open up an issue."

Motivation

I'd like to run LLaVa on AWS Inferentia.

Your contribution

I can help with testing the eventual implementation.

@GrigoryEvko
Copy link

+1, would absolutely love running llava on trn1 and inf2

@lifo9
Copy link
Author

lifo9 commented Mar 27, 2024

Bump

@cszhz
Copy link

cszhz commented Apr 27, 2024

+1

@swy-bys
Copy link

swy-bys commented May 23, 2024

It would be great to add a VLM into the supported models.

@cszhz
Copy link

cszhz commented Sep 24, 2024

Copy link

This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 5 days.

@github-actions github-actions bot added the Stale label Oct 24, 2024
Copy link

This issue was closed because it has been stalled for 5 days with no activity.

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Oct 30, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

5 participants