Deploy ChatQnA on Kubernetes cluster¶
You should have Helm (version >= 3.15) installed. Refer to the Helm Installation Guide for more information.
For more deploy options, refer to helm charts README.
Deploy on Xeon¶
export HFTOKEN="insert-your-huggingface-token-here"
helm install chatqna oci://ghcr.io/opea-project/charts/chatqna --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} -f cpu-values.yaml
Deploy on Gaudi¶
export HFTOKEN="insert-your-huggingface-token-here"
helm install chatqna oci://ghcr.io/opea-project/charts/chatqna --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} -f gaudi-values.yaml
Deploy variants of ChatQnA¶
ChatQnA is configurable and you can enable/disable features by providing values.yaml file. For example, to run with vllm instead of tgi on Gaudi hardware, use gaudi-vllm-values.yaml file:
export HFTOKEN="insert-your-huggingface-token-here"
helm install chatqna oci://ghcr.io/opea-project/charts/chatqna --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} -f gaudi-vllm-values.yaml
See other *-values.yaml files in this directory for more reference.