Replies: 2 comments
-
Hey @gooseillo don't use inference - but only "preprocessing". Have a look at how we build our lightweight Docker image |
Beta Was this translation helpful? Give feedback.
0 replies
-
Perhaps it would be useful to see how we make Docker images at https://github.com/deepset-ai/haystack/tree/main/docker and pay attention to cpu-remote-inference in our bake file |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi everyone,
I'm trying to deploy a haystack RAG pipeline to EBS or EC2 on one of the lower config machines (2CPU - 2GB RAM and 2CPU 4GB RAM). While building the container, I used pip install 'farm-haystack[inference]' ## installs torch, sentence-transformers, sentencepiece, and huggingface-hub.
I was wondering, since I only intend to do inference on the CPU, why does it have to install torch and other heavier libraries. Is there a lighter install?
Otherwise, is there a cheaper deployment pipeline for a RAG solution?
Beta Was this translation helpful? Give feedback.
All reactions