agent¶
Helm chart for deploying Agent microservice.
agent depends on LLM service, you should set llm_endpoint_url as LLM endpoint.
Deploy¶
Use external LLM endpoint¶
helm install agent oci://ghcr.io/opea-project/charts/agent –set llm_endpoint_url=${YOUR_LLM_ENDPOINT}
Deploy with tgi¶
helm install agent oci://ghcr.io/opea-project/charts/agent –set tgi.enabled=True
Deploy with vllm¶
helm install agent oci://ghcr.io/opea-project/charts/agent –set vllm.enabled=True
Verify¶
To verify the installation, run the command kubectl get pod
to make sure all pods are running.
Then run the command kubectl port-forward svc/agent 9090:9090
to expose the agent service for access.
Open another terminal and run the following command to verify the service if working:
curl http://localhost:9090/v1/chat/completions \
-X POST \
-H 'Content-Type: application/json' \
-d '{"query":"What is OPEA?"}'
Options¶
For global options, see Global Options.
Key |
Type |
Default |
Description |
---|---|---|---|
global.HUGGINGFACEHUB_API_TOKEN |
string |
|
Your own Hugging Face API token |
image.repository |
string |
|
|
service.port |
string |
|
|
llm_endpoint_url |
string |
|
LLM endpoint |
global.monitoring |
bop; |
false |
Service usage metrics |