Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

video-salmonn vicuna question #86

Open
HeChengHui opened this issue Nov 29, 2024 · 2 comments
Open

video-salmonn vicuna question #86

HeChengHui opened this issue Nov 29, 2024 · 2 comments

Comments

@HeChengHui
Copy link

@BriansIDP
Thank you for your work.

what is the vram requirement to run inference? i am having OOM using lmsys/vicuna-13b, but lmsys/vicuna-7b is giving me size mismatch error.
Or am i using the wrong model?

@BriansIDP
Copy link
Collaborator

Thank you for the question.
Video-SALMONN is trained with vicuna-13b so the input dimension for the 7b model would not match the Q-Former output of video-SALMONN. It would be helpful to try quantization (with a bit of performance loss).

@HeChengHui
Copy link
Author

HeChengHui commented Nov 29, 2024

@BriansIDP
does that mean i can use something like TheBloke/vicuna-13B-v1.5-16K-AWQ by just setting it in the config?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants