Skip to content

Commit 8cea925

Browse files
committed
update: address review comments
- revert back to use llama3.2:1b - remove unnecessary/unrelated comments/changes - set INFERNECE_MODEL to OLLAMA_INFERENCE_MODEL - remove ENALBE_OLLAMA - set images to use "latest" tag than 0.2.15 - Signed-off-by: Wen Zhou <[email protected]>
1 parent e10138d commit 8cea925

File tree

6 files changed

+14
-27
lines changed

6 files changed

+14
-27
lines changed

README.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -84,18 +84,16 @@ spec:
8484
containerSpec:
8585
env:
8686
- name: INFERENCE_MODEL
87-
value: "llama3.2:3b"
87+
value: "llama3.2:1b"
8888
- name: OLLAMA_URL
8989
value: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
90-
- name: ENABLE_OLLAMA
91-
value: ollama
9290
storage:
9391
size: "20Gi"
9492
mountPath: "/home/lls/.lls"
9593
```
9694
3. Verify the server pod is running in the user defined namespace.
9795

98-
### Using a ConfigMap to override default run.yaml configuration from distribution
96+
### Using a ConfigMap for run.yaml configuration
9997

10098
A ConfigMap can be used to store run.yaml configuration for each LlamaStackDistribution.
10199
Updates to the ConfigMap will restart the Pod to load the new data.

config/manager/manager.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ spec:
4848
allowPrivilegeEscalation: false
4949
capabilities:
5050
drop:
51-
- "ALL"
51+
- "ALL"
5252
livenessProbe:
5353
httpGet:
5454
path: /healthz

config/samples/_v1alpha1_llamastackdistribution.yaml

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -7,12 +7,10 @@ spec:
77
server:
88
containerSpec:
99
env:
10-
- name: INFERENCE_MODEL
11-
value: 'llama3.2:3b'
10+
- name: OLLAMA_INFERENCE_MODEL
11+
value: 'llama3.2:1b'
1212
- name: OLLAMA_URL
1313
value: 'http://ollama-server-service.ollama-dist.svc.cluster.local:11434'
14-
- name: ENABLE_OLLAMA
15-
value: ollama
1614
name: llama-stack
1715
distribution:
1816
name: starter

config/samples/example-with-configmap.yaml

Lines changed: 4 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -11,20 +11,15 @@ data:
1111
- inference
1212
providers:
1313
inference:
14-
- provider_id: ${env.ENABLE_OLLAMA:=__disabled__}
14+
- provider_id: ollama
1515
provider_type: "remote::ollama"
1616
config:
1717
url: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
1818
models:
19-
- model_id: "ollama/llama3.2:3b"
19+
- model_id: "llama3.2:1b"
2020
provider_id: ollama
2121
model_type: llm
22-
provider_model_id: llama3.2:3b
23-
- embedding_dimension: 384
24-
model_id: ${env.ENABLE_OLLAMA:=__disabled__}/${env.OLLAMA_EMBEDDING_MODEL:=__disabled__}
25-
provider_id: ${env.ENABLE_OLLAMA:=__disabled__}
26-
provider_model_id: ${env.OLLAMA_EMBEDDING_MODEL:=__disabled__}
27-
model_type: embedding
22+
provider_model_id: llama3.2:1b
2823
server:
2924
port: 8321
3025
---
@@ -40,10 +35,8 @@ spec:
4035
containerSpec:
4136
port: 8321
4237
env:
43-
- name: ENABLE_OLLAMA
44-
value: ollama
4538
- name: OLLAMA_EMBEDDING_MODEL
4639
value: all-minilm:l6-v2
4740
userConfig:
48-
configMapName: llama-stack-config # use ConfigMap's data.run.yaml
41+
configMapName: llama-stack-config
4942
# configMapNamespace: "" # Optional - defaults to the same namespace as the CR

config/samples/example-withoutconfigmpa.yaml

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,11 +12,9 @@ spec:
1212
port: 8321
1313
env:
1414
- name: OLLAMA_INFERENCE_MODEL
15-
value: "llama3.2:3b"
15+
value: "llama3.2:1b"
1616
- name: OLLAMA_URL
1717
value: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
18-
- name: ENABLE_OLLAMA
19-
value: ollama
2018
storage:
2119
size: "10Gi" # Optional - defaults to 10Gi
2220
mountPath: "/home/lls/.lls" # Optional - defaults to /.llama

distributions.json

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
{
2-
"starter": "docker.io/llamastack/distribution-starter:0.2.15",
3-
"remote-vllm": "docker.io/llamastack/distribution-remote-vllm:0.2.15",
4-
"meta-reference-gpu": "docker.io/llamastack/distribution-meta-reference-gpu:0.2.15",
5-
"postgres-demo": "docker.io/llamastack/distribution-postgres-demo:0.2.15"
2+
"starter": "docker.io/llamastack/distribution-starter:latest",
3+
"remote-vllm": "docker.io/llamastack/distribution-remote-vllm:latest",
4+
"meta-reference-gpu": "docker.io/llamastack/distribution-meta-reference-gpu:latest",
5+
"postgres-demo": "docker.io/llamastack/distribution-postgres-demo:latest"
66
}

0 commit comments

Comments
 (0)