Skip to content
This repository has been archived by the owner on Mar 30, 2024. It is now read-only.

Commit

Permalink
Pass TRUNCATE_PROMPT_LENGTH to deployment, switch to ghcr.io image (
Browse files Browse the repository at this point in the history
#63)

Signed-off-by: Hung-Han (Henry) Chen <chenhungh@gmail.com>
  • Loading branch information
chenhunghan authored Aug 26, 2023
1 parent 2f03dee commit 4f651e3
Show file tree
Hide file tree
Showing 3 changed files with 5 additions and 3 deletions.
4 changes: 2 additions & 2 deletions charts/ialacol/Chart.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
apiVersion: v2
appVersion: 0.11.0
appVersion: 0.11.1
description: A Helm chart for ialacol
name: ialacol
type: application
version: 0.11.0
version: 0.11.1
2 changes: 2 additions & 0 deletions charts/ialacol/templates/deployment.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -57,6 +57,8 @@ spec:
value: {{ (.Values.deployment.env).GPU_LAYERS | quote }}
- name: MODE_TYPE
value: {{ (.Values.deployment.env).MODE_TYPE | quote }}
- name: TRUNCATE_PROMPT_LENGTH
value: {{ (.Values.deployment.env).TRUNCATE_PROMPT_LENGTH | quote }}
volumeMounts:
- mountPath: /app/models
name: model
Expand Down
2 changes: 1 addition & 1 deletion charts/ialacol/values.yaml
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
replicas: 1

deployment:
image: quay.io/chenhunghan/ialacol:latest
image: ghcr.io/chenhunghan/ialacol:latest
# or use CUDA image `ghcr.io/chenhunghan/ialacol-cuda12:latest`
# env:
# DEFAULT_MODEL_HG_REPO_ID: TheBloke/Llama-2-7B-Chat-GGML
Expand Down

0 comments on commit 4f651e3

Please sign in to comment.