Best way to run GPT4-x-Alpaca-13B on 3060TI / 16GB ram? #749
Unanswered
Nightnightlight
asked this question in
Q&A
Replies: 2 comments 3 replies
-
you sure you're using GPU inference? llama.cpp-based (ggml) models run on CPU. |
Beta Was this translation helpful? Give feedback.
2 replies
-
using |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I can just barely run it with default settings from the installer at 1.09 tokens a second. Anyway to perhaps improve those speeds or is that the best my setup is gonna get?
Beta Was this translation helpful? Give feedback.
All reactions