I need help to quantize video-llava model to quant size 4 or 8 #6029
dnabanita7
started this conversation in
General
Replies: 1 comment
-
you can use bitsandbytes to convert the model, i assume it is an easiest way load the model from transformer |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am trying to convert video-llava according to instructions at #2948 but I am having issues because of
convert.py
file. Since, this is llava + vicuna should I add anything extra in the convert.py file or let it remain the same?Beta Was this translation helpful? Give feedback.
All reactions