# Build and deploy CodeGen Application on AMD GPU (ROCm) ## Build images ### Build the LLM Docker Image ```bash ### Cloning repo git clone https://github.com/opea-project/GenAIComps.git cd GenAIComps ### Build Docker image docker build -t opea/llm-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/llms/text-generation/tgi/Dockerfile . ``` ### Build the MegaService Docker Image ```bash ### Cloning repo git clone https://github.com/opea-project/GenAIExamples cd GenAIExamples/CodeGen ### Build Docker image docker build -t opea/codegen:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile . ``` ### Build the UI Docker Image ```bash cd GenAIExamples/CodeGen/ui ### Build UI Docker image docker build -t opea/codegen-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile . ### Build React UI Docker image (React UI allows you to use file uploads) docker build --no-cache -t opea/codegen-react-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile.react . ``` It is recommended to use the React UI as it works for downloading files. The use of React UI is set in the Docker Compose file ## Deploy CodeGen Application ### Features of Docker compose for AMD GPUs 1. Added forwarding of GPU devices to the container TGI service with instructions: ```yaml shm_size: 1g devices: - /dev/kfd:/dev/kfd - /dev/dri/:/dev/dri/ cap_add: - SYS_PTRACE group_add: - video security_opt: - seccomp:unconfined ``` In this case, all GPUs are thrown. To reset a specific GPU, you need to use specific device names cardN and renderN. For example: ```yaml shm_size: 1g devices: - /dev/kfd:/dev/kfd - /dev/dri/card0:/dev/dri/card0 - /dev/dri/render128:/dev/dri/render128 cap_add: - SYS_PTRACE group_add: - video security_opt: - seccomp:unconfined ``` To find out which GPU device IDs cardN and renderN correspond to the same GPU, use the GPU driver utility ### Go to the directory with the Docker compose file ```bash cd GenAIExamples/CodeGen/docker_compose/amd/gpu/rocm ``` ### Set environments In the file "GenAIExamples/CodeGen/docker_compose/amd/gpu/rocm/set_env.sh " it is necessary to set the required values. Parameter assignments are specified in the comments for each variable setting command ```bash chmod +x set_env.sh . set_env.sh ``` ### Run services ``` docker compose up -d ``` # Validate the MicroServices and MegaService ## Validate TGI service ```bash curl http://${HOST_IP}:${CODEGEN_TGI_SERVICE_PORT}/generate \ -X POST \ -d '{"inputs":"Implement a high-level API for a TODO list application. The API takes as input an operation request and updates the TODO list in place. If the request is invalid, raise an exception.","parameters":{"max_new_tokens":256, "do_sample": true}}' \ -H 'Content-Type: application/json' ``` ## Validate LLM service ```bash curl http://${HOST_IP}:${CODEGEN_LLM_SERVICE_PORT}/v1/chat/completions\ -X POST \ -d '{"query":"Implement a high-level API for a TODO list application. The API takes as input an operation request and updates the TODO list in place. If the request is invalid, raise an exception.","max_tokens":256,"top_k":10,"top_p":0.95,"typical_p":0.95,"temperature":0.01,"repetition_penalty":1.03,"streaming":true}' \ -H 'Content-Type: application/json' ``` ## Validate MegaService ```bash curl http://${HOST_IP}:${CODEGEN_BACKEND_SERVICE_PORT}/v1/codegen -H "Content-Type: application/json" -d '{ "messages": "Implement a high-level API for a TODO list application. The API takes as input an operation request and updates the TODO list in place. If the request is invalid, raise an exception." }' ```