Skip to content

Commit

Permalink
fix UI keycloak var, gpu_runtime
Browse files Browse the repository at this point in the history
  • Loading branch information
justinthelaw committed Sep 5, 2024
1 parent 1b4df8e commit 3134908
Show file tree
Hide file tree
Showing 10 changed files with 15 additions and 15 deletions.
4 changes: 2 additions & 2 deletions Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -304,8 +304,8 @@ silent-deploy-gpu:
@$(MAKE) -j${MAX_JOBS} \
silent-deploy-api-package ZARF_FLAGS="${ZARF_FLAGS} ${SILENT_ZARF_FLAGS}" \
silent-deploy-vllm-package ZARF_FLAGS="${ZARF_FLAGS} ${SILENT_ZARF_FLAGS}" \
silent-deploy-text-embeddings-package ZARF_FLAGS="${ZARF_FLAGS} ${SILENT_ZARF_FLAGS} --set=GPU_RUNTIME_NAME='nvidia'" \
silent-deploy-whisper-package ZARF_FLAGS="${ZARF_FLAGS} ${SILENT_ZARF_FLAGS} --set=GPU_RUNTIME_NAME='nvidia'"
silent-deploy-text-embeddings-package ZARF_FLAGS="${ZARF_FLAGS} ${SILENT_ZARF_FLAGS} --set=GPU_RUNTIME='nvidia'" \
silent-deploy-whisper-package ZARF_FLAGS="${ZARF_FLAGS} ${SILENT_ZARF_FLAGS} --set=GPU_RUNTIME='nvidia'"
@echo "Deploying UI..."
@$(MAKE) silent-deploy-ui-package ZARF_FLAGS="${ZARF_FLAGS} ${SILENT_ZARF_FLAGS} --set=MODEL='vllm'"
@echo "All deployments completed"
Expand Down
4 changes: 2 additions & 2 deletions bundles/dev/cpu/uds-config.yaml
Original file line number Diff line number Diff line change
@@ -1,10 +1,10 @@
variables:
text-embeddings:
GPU_RUNTIME_NAME: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_runtime: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_limit: 0

whisper:
GPU_RUNTIME_NAME: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_runtime: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_limit: 0

supabase:
Expand Down
4 changes: 2 additions & 2 deletions bundles/dev/gpu/uds-config.yaml
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
# see individual zarf packaging configuration for more variables and variable descriptions
variables:
text-embeddings:
GPU_RUNTIME_NAME: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_runtime: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_limit: 0 # runs on CPU until GPU limit is increased

whisper:
GPU_RUNTIME_NAME: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_runtime: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_limit: 0 # runs on CPU until GPU limit is increased

vllm:
Expand Down
4 changes: 2 additions & 2 deletions bundles/latest/cpu/uds-config.yaml
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
# see individual zarf packaging configuration for more variables and variable descriptions
variables:
text-embeddings:
GPU_RUNTIME_NAME: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_runtime: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_limit: 0

whisper:
GPU_RUNTIME_NAME: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_runtime: "" # Leave blank if nvidia runtimeClass is not present in cluster
gpu_limit: 0

supabase:
Expand Down
4 changes: 2 additions & 2 deletions bundles/latest/gpu/uds-config.yaml
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
# see individual zarf packaging configuration for more variables and variable descriptions
variables:
text-embeddings:
GPU_RUNTIME_NAME: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_runtime: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_limit: 0 # runs on CPU until GPU limit is increased

whisper:
GPU_RUNTIME_NAME: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_runtime: "nvidia" # Set to ensure the nvidia runtimeClass is present in case GPU limit is increased
gpu_limit: 0 # runs on CPU until GPU limit is increased

vllm:
Expand Down
2 changes: 1 addition & 1 deletion packages/text-embeddings/zarf.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ variables:
description: The GPU limit for the model inferencing.
default: "0"
pattern: "^[0-9]+$"
- name: GPU_RUNTIME_NAME
- name: GPU_RUNTIME
description: The GPU class name for the model inferencing. Leave blank for CPU-only.
default: ""
pattern: "^(nvidia)?$"
Expand Down
2 changes: 1 addition & 1 deletion packages/ui/values/upstream-values.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ env:
- name: PUBLIC_SUPABASE_URL
value: "https://supabase-kong.###ZARF_VAR_DOMAIN###"
- name: PUBLIC_DISABLE_KEYCLOAK
value: "true"
value: "###ZARF_VAR_DISABLE_KEYCLOAK###"
- name: PUBLIC_MESSAGE_LENGTH_LIMIT
value: "10000"
- name: SUPABASE_AUTH_EXTERNAL_KEYCLOAK_URL
Expand Down
2 changes: 1 addition & 1 deletion packages/ui/zarf.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ variables:
prompt: true
sensitive: true
- name: OPENAI_API_KEY
description: OpenAI API Key. If specified, app will use OpenAI instead of LeapfrogAI
description: OpenAI API key - if specified, the UI will use OpenAI instead of LeapfrogAI
prompt: true
default: ""
sensitive: true
Expand Down
2 changes: 1 addition & 1 deletion packages/vllm/zarf.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ variables:
description: The GPU limit for the model inferencing. Must be 1 or more.
default: "1"
pattern: "^[1-9][0-9]*$"
- name: GPU_RUNTIME_NAME
- name: GPU_RUNTIME
description: The GPU runtime name for the model inferencing.
default: "nvidia"
pattern: "^(nvidia)?$"
Expand Down
2 changes: 1 addition & 1 deletion packages/whisper/zarf.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ variables:
description: The GPU limit for the model inferencing.
default: "0"
pattern: "^[0-9]+$"
- name: GPU_RUNTIME_NAME
- name: GPU_RUNTIME
description: The GPU runtime name for the model inferencing. Leave blank for CPU-only.
default: ""
pattern: "^(nvidia)?$"
Expand Down

0 comments on commit 3134908

Please sign in to comment.