File tree Expand file tree Collapse file tree 6 files changed +14
-27
lines changed Expand file tree Collapse file tree 6 files changed +14
-27
lines changed Original file line number Diff line number Diff line change @@ -84,18 +84,16 @@ spec:
8484 containerSpec:
8585 env:
8686 - name: INFERENCE_MODEL
87- value: "llama3.2:3b "
87+ value: "llama3.2:1b "
8888 - name: OLLAMA_URL
8989 value: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
90- - name: ENABLE_OLLAMA
91- value: ollama
9290 storage:
9391 size: "20Gi"
9492 mountPath: "/home/lls/.lls"
9593```
96943 . Verify the server pod is running in the user defined namespace.
9795
98- ### Using a ConfigMap to override default run.yaml configuration from distribution
96+ ### Using a ConfigMap for run.yaml configuration
9997
10098A ConfigMap can be used to store run.yaml configuration for each LlamaStackDistribution.
10199Updates to the ConfigMap will restart the Pod to load the new data.
Original file line number Diff line number Diff line change 4848 allowPrivilegeEscalation : false
4949 capabilities :
5050 drop :
51- - " ALL"
51+ - " ALL"
5252 livenessProbe :
5353 httpGet :
5454 path : /healthz
Original file line number Diff line number Diff line change 77 server :
88 containerSpec :
99 env :
10- - name : INFERENCE_MODEL
11- value : ' llama3.2:3b '
10+ - name : OLLAMA_INFERENCE_MODEL
11+ value : ' llama3.2:1b '
1212 - name : OLLAMA_URL
1313 value : ' http://ollama-server-service.ollama-dist.svc.cluster.local:11434'
14- - name : ENABLE_OLLAMA
15- value : ollama
1614 name : llama-stack
1715 distribution :
1816 name : starter
Original file line number Diff line number Diff line change @@ -11,20 +11,15 @@ data:
1111 - inference
1212 providers:
1313 inference:
14- - provider_id: ${env.ENABLE_OLLAMA:=__disabled__}
14+ - provider_id: ollama
1515 provider_type: "remote::ollama"
1616 config:
1717 url: "http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
1818 models:
19- - model_id: "ollama/ llama3.2:3b "
19+ - model_id: "llama3.2:1b "
2020 provider_id: ollama
2121 model_type: llm
22- provider_model_id: llama3.2:3b
23- - embedding_dimension: 384
24- model_id: ${env.ENABLE_OLLAMA:=__disabled__}/${env.OLLAMA_EMBEDDING_MODEL:=__disabled__}
25- provider_id: ${env.ENABLE_OLLAMA:=__disabled__}
26- provider_model_id: ${env.OLLAMA_EMBEDDING_MODEL:=__disabled__}
27- model_type: embedding
22+ provider_model_id: llama3.2:1b
2823 server:
2924 port: 8321
3025 ---
4035 containerSpec :
4136 port : 8321
4237 env :
43- - name : ENABLE_OLLAMA
44- value : ollama
4538 - name : OLLAMA_EMBEDDING_MODEL
4639 value : all-minilm:l6-v2
4740 userConfig :
48- configMapName : llama-stack-config # use ConfigMap's data.run.yaml
41+ configMapName : llama-stack-config
4942 # configMapNamespace: "" # Optional - defaults to the same namespace as the CR
Original file line number Diff line number Diff line change 1212 port : 8321
1313 env :
1414 - name : OLLAMA_INFERENCE_MODEL
15- value : " llama3.2:3b "
15+ value : " llama3.2:1b "
1616 - name : OLLAMA_URL
1717 value : " http://ollama-server-service.ollama-dist.svc.cluster.local:11434"
18- - name : ENABLE_OLLAMA
19- value : ollama
2018 storage :
2119 size : " 10Gi" # Optional - defaults to 10Gi
2220 mountPath : " /home/lls/.lls" # Optional - defaults to /.llama
Original file line number Diff line number Diff line change 11{
2- "starter" : " docker.io/llamastack/distribution-starter:0.2.15 " ,
3- "remote-vllm" : " docker.io/llamastack/distribution-remote-vllm:0.2.15 " ,
4- "meta-reference-gpu" : " docker.io/llamastack/distribution-meta-reference-gpu:0.2.15 " ,
5- "postgres-demo" : " docker.io/llamastack/distribution-postgres-demo:0.2.15 "
2+ "starter" : " docker.io/llamastack/distribution-starter:latest " ,
3+ "remote-vllm" : " docker.io/llamastack/distribution-remote-vllm:latest " ,
4+ "meta-reference-gpu" : " docker.io/llamastack/distribution-meta-reference-gpu:latest " ,
5+ "postgres-demo" : " docker.io/llamastack/distribution-postgres-demo:latest "
66}
You can’t perform that action at this time.
0 commit comments