-
Notifications
You must be signed in to change notification settings - Fork 1
/
script
54 lines (40 loc) · 1.66 KB
/
script
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
kubectl exec -it deploy/client -- curl http://ollama.ollama:80/api/generate -d '{"model": "llama3.2", "prompt":"what are the top 1 favorite place in Salt Lake City?", "stream": false}'
for i in {2..10009}
do
kubectl exec -it deploy/client -- curl http://ollama.ollama:80/api/tags
sleep 5
done
for i in {2..10009}
do
kubectl exec -it deploy/client -- curl http://ollama.ollama:80/api/generate -d '{"model": "llama3.2", "prompt":"what are the top 1 favorite place in Salt Lake City?", "stream": false}'
done
kubectl port-forward -n ollama service/ollama 11434:80
kubectl port-forward service/demo 8501:8501
ollama pull llama3.2
ollama pull llava
curl http://localhost:11434/api/generate -d '{
"model": "llama3.2",
"prompt": "what are the top 2 favorite places in Salt Lake City?",
"stream": false
}'
kubectl exec -it deploy/client -- curl http://ollama.ollama:80/api/pull -d '{"name": "llama3.2"}'
kubectl exec -it deploy/client -- curl http://ollama.ollama:80/api/pull -d '{"name": "llava"}'
curl http://localhost:11434/api/chat -d '{
"model": "llama3.2",
"messages": [
{"role": "user", "content": "what are top 2 favorite places in Salt Lake City?"}
],
"stream": false
}'
kubectl exec -it deploy/client -- curl http://ollama.ollama:80/api/chat -d '{"model": "llama3.2", "messages": [{"role": "user", "content": "what are top 2 favorite places in Salt Lake City?"}], "stream": false}'
curl http://localhost:11434/api/chat -d '{
"model": "llama3.2",
"messages": [
{"role": "user", "content": "what are top 2 favorite places in Salt Lake City?"}
],
"stream": true
}'
for i in {2..10009}
do
k exec -it deploy/client -- curl http://httpbin.org/get -v
done