Skip to content

Commit

Permalink
prep to have a cpu and gpu version
Browse files Browse the repository at this point in the history
  • Loading branch information
Schnitzel committed Aug 17, 2023
1 parent d49f6b3 commit a697ed6
Show file tree
Hide file tree
Showing 4 changed files with 9 additions and 7 deletions.
7 changes: 1 addition & 6 deletions .lagoon.yml
Original file line number Diff line number Diff line change
@@ -1,6 +1 @@
docker-compose-yaml: docker-compose.lagoon.yml
environments:
main:
routes:
- llama2-api:
- llama2.marco-test6.amazeeio.review
docker-compose-yaml: docker-compose.gpu.yml
File renamed without changes.
7 changes: 7 additions & 0 deletions docker-compose.yml
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,11 @@ services:
build:
context: .
dockerfile: Dockerfile
labels:
lagoon.type: basic-persistent
lagoon.persistent: /data
lagoon.service.port: 8000
lagoon.autogeneratedroute: true
volumes:
- ./data:/data:delegated
ports:
Expand All @@ -14,5 +19,7 @@ services:
build:
context: .
dockerfile: ui.Dockerfile
labels:
lagoon.type: node
ports:
- "3000:3000"
2 changes: 1 addition & 1 deletion start-llama2.sh
Original file line number Diff line number Diff line change
Expand Up @@ -5,4 +5,4 @@ ulimit -l unlimited
echo "Model: ${MODEL}"

python3 hug_model.py -s ${MODEL} -f "q5_1"
python3 -B -m llama_cpp.server --model /data/model.bin --n_gpu_layers=43
python3 -B -m llama_cpp.server --model /data/model.bin ${LLAMA_CPP_ARGS}

0 comments on commit a697ed6

Please sign in to comment.