File tree Expand file tree Collapse file tree 5 files changed +21
-38
lines changed Expand file tree Collapse file tree 5 files changed +21
-38
lines changed Original file line number Diff line number Diff line change @@ -42,7 +42,7 @@ kubectl apply -f https://raw.githubusercontent.com/llamastack/llama-stack-k8s-op
4242
4343### Deploying the Llama Stack Server
4444
45- 1 . Deploy the inference provider server (ollama, vllm etc ). Example to deploy a new ollama server:
45+ 1 . Deploy the inference provider server (ollama). Example to deploy a new ollama server:
4646```
4747bash hack/deploy-ollama.sh
4848```
@@ -56,14 +56,15 @@ spec:
5656 replicas: 1
5757 server:
5858 distribution:
59- name: ollama
59+ name: starter
6060 containerSpec:
61- port: 8321
6261 env:
6362 - name: INFERENCE_MODEL
64- value: "llama3.2:1b "
63+ value: "llama3.2:3b "
6564 - name: OLLAMA_URL
6665 value: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
66+ - name: ENABLE_OLLAMA
67+ value: ollama
6768 storage:
6869 size: "20Gi"
6970 mountPath: "/home/lls/.lls"
Original file line number Diff line number Diff line change @@ -8,18 +8,7 @@ metadata:
88# The distribution images are immutable and should not be updated
99immutable : true
1010data :
11- ollama : docker.io/llamastack/distribution-ollama:latest
12- hf-endpoint : docker.io/llamastack/distribution-hf-endpoint:latest
13- hf-serverless : docker.io/llamastack/distribution-hf-serverless:latest
14- bedrock : docker.io/llamastack/distribution-bedrock:latest
15- cerebras : docker.io/llamastack/distribution-cerebras:latest
16- nvidia : docker.io/llamastack/distribution-nvidia:latest
17- open-benchmark : docker.io/llamastack/distribution-open-benchmark:latest
18- passthrough : docker.io/llamastack/distribution-passthrough:latest
19- remote-vllm : docker.io/llamastack/distribution-remote-vllm:latest
20- sambanova : docker.io/llamastack/distribution-sambanova:latest
11+ meta-reference-gpu : docker.io/llamastack/distribution-meta-reference-gpu:latest
2112 tgi : docker.io/llamastack/distribution-tgi:latest
22- together : docker.io/llamastack/distribution-together:latest
23- vllm-gpu : docker.io/llamastack/distribution-vllm-gpu:latest
24- watsonx : docker.io/llamastack/distribution-watsonx:latest
25- fireworks : docker.io/llamastack/distribution-fireworks:latest
13+ postgres-demo : docker.io/llamastack/distribution-postgres-demo:latest
14+ starter : docker.io/llamastack/distribution-starter:latest
Original file line number Diff line number Diff line change 88 containerSpec :
99 env :
1010 - name : INFERENCE_MODEL
11- value : ' llama3.2:1b '
11+ value : ' llama3.2:3b '
1212 - name : OLLAMA_URL
1313 value : ' http://ollama-server-service.ollama-dist.svc.cluster.local:11434'
14+ - name : ENABLE_OLLAMA
15+ value : ollama
1416 name : llama-stack
1517 distribution :
16- name : ollama
18+ name : starter
1719 # Uncomment the storage section to use persistent storage
1820 # storage: {} # Will use default size of 10Gi and default mount path of /.llama
1921 # Or specify custom values:
Original file line number Diff line number Diff line change 66 run.yaml : |
77 # Llama Stack Configuration
88 version: '2'
9- image_name: ollama
9+ image_name: starter
1010 apis:
1111 - inference
1212 providers:
1616 config:
1717 url: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
1818 models:
19- - model_id: "llama3.2:1b "
19+ - model_id: "ollama/ llama3.2:3b "
2020 provider_id: ollama
2121 model_type: llm
2222 server:
@@ -30,14 +30,16 @@ spec:
3030 replicas : 1
3131 server :
3232 distribution :
33- name : ollama
33+ name : starter
3434 containerSpec :
3535 port : 8321
3636 env :
3737 - name : INFERENCE_MODEL
38- value : " llama3.2:1b "
38+ value : " llama3.2:3b "
3939 - name : OLLAMA_URL
4040 value : " http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
41+ - name : ENABLE_OLLAMA
42+ value : ollama
4143 userConfig :
4244 configMapName : llama-stack-config
4345 # configMapNamespace: "" # Optional - defaults to the same namespace as the CR
Original file line number Diff line number Diff line change @@ -2400,21 +2400,10 @@ subjects:
24002400---
24012401apiVersion : v1
24022402data :
2403- bedrock : docker.io/llamastack/distribution-bedrock:latest
2404- cerebras : docker.io/llamastack/distribution-cerebras:latest
2405- fireworks : docker.io/llamastack/distribution-fireworks:latest
2406- hf-endpoint : docker.io/llamastack/distribution-hf-endpoint:latest
2407- hf-serverless : docker.io/llamastack/distribution-hf-serverless:latest
2408- nvidia : docker.io/llamastack/distribution-nvidia:latest
2409- ollama : docker.io/llamastack/distribution-ollama:latest
2410- open-benchmark : docker.io/llamastack/distribution-open-benchmark:latest
2411- passthrough : docker.io/llamastack/distribution-passthrough:latest
2412- remote-vllm : docker.io/llamastack/distribution-remote-vllm:latest
2413- sambanova : docker.io/llamastack/distribution-sambanova:latest
2403+ meta-reference-gpu : docker.io/llamastack/distribution-meta-reference-gpu:latest
2404+ postgres-demo : docker.io/llamastack/distribution-postgres-demo:latest
2405+ starter : docker.io/llamastack/distribution-starter:latest
24142406 tgi : docker.io/llamastack/distribution-tgi:latest
2415- together : docker.io/llamastack/distribution-together:latest
2416- vllm-gpu : docker.io/llamastack/distribution-vllm-gpu:latest
2417- watsonx : docker.io/llamastack/distribution-watsonx:latest
24182407immutable : true
24192408kind : ConfigMap
24202409metadata :
You can’t perform that action at this time.
0 commit comments