# agent Helm chart for deploying Agent microservice. agent depends on LLM service, you should set llm_endpoint_url as LLM endpoint. ## Deploy ### Use external LLM endpoint helm install agent oci://ghcr.io/opea-project/charts/agent --set llm_endpoint_url=${YOUR_LLM_ENDPOINT} ### Deploy with tgi helm install agent oci://ghcr.io/opea-project/charts/agent --set tgi.enabled=True ### Deploy with vllm helm install agent oci://ghcr.io/opea-project/charts/agent --set vllm.enabled=True ## Verify To verify the installation, run the command `kubectl get pod` to make sure all pods are running. Then run the command `kubectl port-forward svc/agent 9090:9090` to expose the agent service for access. Open another terminal and run the following command to verify the service if working: ```console curl http://localhost:9090/v1/chat/completions \ -X POST \ -H 'Content-Type: application/json' \ -d '{"query":"What is OPEA?"}' ``` ## Options For global options, see Global Options. | Key | Type | Default | Description | | ------------------------------- | ------ | -------------- | --------------------------------------------------------------------------------------- | | global.HUGGINGFACEHUB_API_TOKEN | string | `""` | Your own Hugging Face API token | | image.repository | string | `"opea/agent"` | | | service.port | string | `"9090"` | | | llm_endpoint_url | string | `""` | LLM endpoint | | toolHostPath | string | `""` | hostPath to be mounted to agent's /home/user/tools, used for passing files for tools | | toolPVC | string | `""` | Same as toolHostPath, but use PVC. You can only specify one of toolHostPath and toolPVC | | global.monitoring | bop; | false | Service usage metrics |