doc_embedder raised 'AssertionError: Torch not compiled with CUDA enabled' #6410
-
Hi team, I tried to deploy: on a t4g.xlarge and got "doc_embedder raised 'AssertionError: Torch not compiled with CUDA enabled'" when executing:
Checked google, discord and this forum with no success. Consider now upgrading to a p2.xlarge - any thoughts? Are there any recommendation to get started with haystack using an AMAZON EC2 instance? Thanks in advance for support and advise. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Hi @keebeegee the example code that you shared uses 4bit quantization, which is only available with CUDA installed. That's why you are seeing this error. Almost all other parts of Haystack do not require CUDA and while they benefit from running on a GPU you can run them just on a CPU too. |
Beta Was this translation helpful? Give feedback.
Hi @keebeegee the example code that you shared uses 4bit quantization, which is only available with CUDA installed. That's why you are seeing this error. Almost all other parts of Haystack do not require CUDA and while they benefit from running on a GPU you can run them just on a CPU too.
If you want to try the open source LLM from the example, I suggest you use the 4bit quantization and switch to an EC2 instance with a GPU and a torch version with CUDA enabled, yes.