From af6310f0e172f9d6fedf201f52d093965901d508 Mon Sep 17 00:00:00 2001 From: sd109 Date: Tue, 5 Nov 2024 14:21:31 +0000 Subject: [PATCH] Update default config for Zenith services --- charts/azimuth-chat/values.yaml | 3 +++ charts/azimuth-image-analysis/values.yaml | 4 ++++ charts/azimuth-llm/templates/ui/ui-zenith-reservation.yml | 2 +- charts/azimuth-llm/values.yaml | 4 ++-- 4 files changed, 10 insertions(+), 3 deletions(-) diff --git a/charts/azimuth-chat/values.yaml b/charts/azimuth-chat/values.yaml index 9a17317..2565bc6 100644 --- a/charts/azimuth-chat/values.yaml +++ b/charts/azimuth-chat/values.yaml @@ -4,6 +4,9 @@ azimuth-llm: ui: image: repository: ghcr.io/stackhpc/azimuth-llm-chat-ui + service: + zenith: + enabled: true appSettings: model_name: *model-name model_instruction: "You are a helpful AI assistant; please respond appropriately." diff --git a/charts/azimuth-image-analysis/values.yaml b/charts/azimuth-image-analysis/values.yaml index 7797de6..804e3ea 100644 --- a/charts/azimuth-image-analysis/values.yaml +++ b/charts/azimuth-image-analysis/values.yaml @@ -4,5 +4,9 @@ azimuth-llm: ui: image: repository: ghcr.io/stackhpc/azimuth-llm-image-analysis-ui + service: + zenith: + enabled: true + label: Image Analysis Interface appSettings: model_name: *model-name diff --git a/charts/azimuth-llm/templates/ui/ui-zenith-reservation.yml b/charts/azimuth-llm/templates/ui/ui-zenith-reservation.yml index c114e76..be2f1a8 100644 --- a/charts/azimuth-llm/templates/ui/ui-zenith-reservation.yml +++ b/charts/azimuth-llm/templates/ui/ui-zenith-reservation.yml @@ -15,4 +15,4 @@ metadata: spec: credentialSecretName: {{ .Release.Name }}-ui-zenith-credential {{- end -}} -{{- end -}} \ No newline at end of file +{{- end -}} diff --git a/charts/azimuth-llm/values.yaml b/charts/azimuth-llm/values.yaml index 253056c..29879d4 100644 --- a/charts/azimuth-llm/values.yaml +++ b/charts/azimuth-llm/values.yaml @@ -46,10 +46,10 @@ api: zenith: enabled: false skipAuth: false - label: Inference API + label: OpenAI API iconUrl: https://raw.githubusercontent.com/vllm-project/vllm/v0.2.7/docs/source/assets/logos/vllm-logo-only-light.png description: | - The raw inference API endpoints for the deployed LLM. + The OpenAI API for the deployed LLM. # Config for huggingface model cache volume # This is mounted at /root/.cache/huggingface in the api deployment cacheVolume: