Replies: 2 comments 9 replies
-
You need to read and learn more about each parameters. |
Beta Was this translation helpful? Give feedback.
-
I got it working with the following options: --auto-devices --chat --model gpt4-x-alpaca-13b-native-4bit-128g --wbits 4 --groupsize 128 --gpu-memory 5 --pre_layer 20 It's a bit slow and dumb though |
Beta Was this translation helpful? Give feedback.
-
Hey,
i have a rtx 3070 ti with 8 gb vram and i tried to just run the WebUI but i always run out of memory before i even can choose which model i want to use. I tried multiple starting options like: --auto-devices --chat --wbits 4 --groupsize 128 --load-in-8bit --no-cache --disk --gpu-memory 8, but nothing seems to work. Has anyone a tip for me?
Beta Was this translation helpful? Give feedback.
All reactions