Skip to content

Commit f029bff

Browse files
committed
docs: update using "starter" distro than "ollama"
- update example - set new env to use ollama - use the same model as in llama-stack - remove deprecated distro images Signed-off-by: Wen Zhou <[email protected]>
1 parent bc93353 commit f029bff

File tree

5 files changed

+21
-38
lines changed

5 files changed

+21
-38
lines changed

README.md

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ kubectl apply -f https://raw.githubusercontent.com/llamastack/llama-stack-k8s-op
4242

4343
### Deploying the Llama Stack Server
4444

45-
1. Deploy the inference provider server (ollama, vllm etc). Example to deploy a new ollama server:
45+
1. Deploy the inference provider server (ollama). Example to deploy a new ollama server:
4646
```
4747
bash hack/deploy-ollama.sh
4848
```
@@ -56,14 +56,15 @@ spec:
5656
replicas: 1
5757
server:
5858
distribution:
59-
name: ollama
59+
name: starter
6060
containerSpec:
61-
port: 8321
6261
env:
6362
- name: INFERENCE_MODEL
64-
value: "llama3.2:1b"
63+
value: "llama3.2:3b"
6564
- name: OLLAMA_URL
6665
value: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
66+
- name: ENABLE_OLLAMA
67+
value: ollama
6768
storage:
6869
size: "20Gi"
6970
mountPath: "/home/lls/.lls"

config/manager/distribution-configmap.yaml

Lines changed: 3 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -8,18 +8,7 @@ metadata:
88
# The distribution images are immutable and should not be updated
99
immutable: true
1010
data:
11-
ollama: docker.io/llamastack/distribution-ollama:latest
12-
hf-endpoint: docker.io/llamastack/distribution-hf-endpoint:latest
13-
hf-serverless: docker.io/llamastack/distribution-hf-serverless:latest
14-
bedrock: docker.io/llamastack/distribution-bedrock:latest
15-
cerebras: docker.io/llamastack/distribution-cerebras:latest
16-
nvidia: docker.io/llamastack/distribution-nvidia:latest
17-
open-benchmark: docker.io/llamastack/distribution-open-benchmark:latest
18-
passthrough: docker.io/llamastack/distribution-passthrough:latest
19-
remote-vllm: docker.io/llamastack/distribution-remote-vllm:latest
20-
sambanova: docker.io/llamastack/distribution-sambanova:latest
11+
meta-reference-gpu: docker.io/llamastack/distribution-meta-reference-gpu:latest
2112
tgi: docker.io/llamastack/distribution-tgi:latest
22-
together: docker.io/llamastack/distribution-together:latest
23-
vllm-gpu: docker.io/llamastack/distribution-vllm-gpu:latest
24-
watsonx: docker.io/llamastack/distribution-watsonx:latest
25-
fireworks: docker.io/llamastack/distribution-fireworks:latest
13+
postgres-demo: docker.io/llamastack/distribution-postgres-demo:latest
14+
starter: docker.io/llamastack/distribution-starter:latest

config/samples/_v1alpha1_llamastackdistribution.yaml

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,12 +8,14 @@ spec:
88
containerSpec:
99
env:
1010
- name: INFERENCE_MODEL
11-
value: 'llama3.2:1b'
11+
value: 'llama3.2:3b'
1212
- name: OLLAMA_URL
1313
value: 'http://ollama-server-service.ollama-dist.svc.cluster.local:11434'
14+
- name: ENABLE_OLLAMA
15+
value: ollama
1416
name: llama-stack
1517
distribution:
16-
name: ollama
18+
name: starter
1719
# Uncomment the storage section to use persistent storage
1820
# storage: {} # Will use default size of 10Gi and default mount path of /.llama
1921
# Or specify custom values:

config/samples/example-with-configmap.yaml

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ data:
66
run.yaml: |
77
# Llama Stack Configuration
88
version: '2'
9-
image_name: ollama
9+
image_name: starter
1010
apis:
1111
- inference
1212
providers:
@@ -16,7 +16,7 @@ data:
1616
config:
1717
url: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
1818
models:
19-
- model_id: "llama3.2:1b"
19+
- model_id: "ollama/llama3.2:3b"
2020
provider_id: ollama
2121
model_type: llm
2222
server:
@@ -30,14 +30,16 @@ spec:
3030
replicas: 1
3131
server:
3232
distribution:
33-
name: ollama
33+
name: starter
3434
containerSpec:
3535
port: 8321
3636
env:
3737
- name: INFERENCE_MODEL
38-
value: "llama3.2:1b"
38+
value: "llama3.2:3b"
3939
- name: OLLAMA_URL
4040
value: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
41+
- name: ENABLE_OLLAMA
42+
value: ollama
4143
userConfig:
4244
configMapName: llama-stack-config
4345
# configMapNamespace: "" # Optional - defaults to the same namespace as the CR

release/operator.yaml

Lines changed: 3 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -2400,21 +2400,10 @@ subjects:
24002400
---
24012401
apiVersion: v1
24022402
data:
2403-
bedrock: docker.io/llamastack/distribution-bedrock:latest
2404-
cerebras: docker.io/llamastack/distribution-cerebras:latest
2405-
fireworks: docker.io/llamastack/distribution-fireworks:latest
2406-
hf-endpoint: docker.io/llamastack/distribution-hf-endpoint:latest
2407-
hf-serverless: docker.io/llamastack/distribution-hf-serverless:latest
2408-
nvidia: docker.io/llamastack/distribution-nvidia:latest
2409-
ollama: docker.io/llamastack/distribution-ollama:latest
2410-
open-benchmark: docker.io/llamastack/distribution-open-benchmark:latest
2411-
passthrough: docker.io/llamastack/distribution-passthrough:latest
2412-
remote-vllm: docker.io/llamastack/distribution-remote-vllm:latest
2413-
sambanova: docker.io/llamastack/distribution-sambanova:latest
2403+
meta-reference-gpu: docker.io/llamastack/distribution-meta-reference-gpu:latest
2404+
postgres-demo: docker.io/llamastack/distribution-postgres-demo:latest
2405+
starter: docker.io/llamastack/distribution-starter:latest
24142406
tgi: docker.io/llamastack/distribution-tgi:latest
2415-
together: docker.io/llamastack/distribution-together:latest
2416-
vllm-gpu: docker.io/llamastack/distribution-vllm-gpu:latest
2417-
watsonx: docker.io/llamastack/distribution-watsonx:latest
24182407
immutable: true
24192408
kind: ConfigMap
24202409
metadata:

0 commit comments

Comments
 (0)