-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llava model quantization seems not be supported #73
Comments
Hey @caojinpei - right now we only support models with I have added supporting vision language models and general |
I am glad to get your reply and thanks for sharing roadmap. Now I want to quantize LLava-v1.6 model whose architecture is LlavaLlamaForCausalLM (Is it XXXForCausalLM?) into W8A16 using GPTQ within llm-compressor. Looking forward your reply, thanks. *Model link: https://huggingface.co/liuhaotian/llava-v1.6-vicuna-7b |
Hi @robertgshaw2-neuralmagic *Model link: https://huggingface.co/llava-hf/llava-v1.6-vicuna-7b-hf |
@caojinpei apologies for the delay, supporting vision-language models is on our roadmap, but not yet supported. We would definitely welcome a PR or an example though! |
Describe the bug
When I use llm-compressor to quantize llava model, but at the begining, it failed. (Unrecognized configuration class: 'transformers.models.llava.configuration_llava.LlavaConfig')
Expected behavior
Hope llm-compressor can support LLaVA model.
To Reproduce
from llmcompressor.transformers import SparseAutoModelForCausalLM, oneshot
MODEL_ID = "/home/models/llava-v1.6-vicuna-7b"
model = SparseAutoModelForCausalLM.from_pretrained(
MODEL_ID,
device_map="auto",
trust_remote_code=True,
)
Errors
ValueError: Unrecognized configuration class <class 'transformers.models.llava.configuration_llava.LlavaConfig'> for this kind of AutoModel
Hope to get your reply, thanks.
The text was updated successfully, but these errors were encountered: