https://www.tecmint.com/install-nvidia-drivers-in-linux/#Method_1_Installing_NVIDIA_Drivers_Using_RPM_Fusion_in_Fedora
=> also install the NVIDIA VAAPI/VDPAU Driver (see end of article)
Cf https://rpmfusion.org/Howto/CUDA#Installation
sudo dnf config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/fedora39/x86_64/cuda-fedora39.repo
sudo dnf clean all
sudo dnf -y install cuda
Cf https://hub.docker.com/r/ollama/ollama Cf https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html#installation Cf https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/cdi-support.html
sudo dnf install -y nvidia-container-toolkit
sudo nvidia-ctk cdi generate --output=/etc/cdi/nvidia.yaml
sudo dnf install -y podman-compose
Define the PGDATA_HOST_DIR and OLLAMA_HOST_DIR variables. either use a .env file or export the variables directly. For instance :
cat >.env <<EOF
PGDATA_HOST_DIR=/data/llm_plgrnd/pgdata
OLLAMA_HOST_DIR=/data/llm_plgrnd/ollama_data
EOF
NB : if not defined, the variables default to "pgdata" and "ollama_data" in the current directory
Start the containers :
mkdir pgdata
podman-compose up -d
Next, import the otrs database and prepare the schema :
psql -h localhost -U postgres -f <dump>
./00_prepare.sh
Generate the vector embeddings :
./01_generate_embeddings.sh
./query_ticket.sh <ticket_number> [-n|--limit <limit>] <-p|--print-conversation>
./query_ticket.sh <your_request> [-n|--limit <limit>] <-p|--print-conversation>
Work in progress...