How to use GPU for inference#
Pre-requisites#
Machine with GPU is available
Configure and deploy GPU pipelines#
In edge-ai-suites/metro-ai-suite/metro-vision-ai-app-recipe/smart-intersection/src/dlstreamer-pipeline-server/config.json the following GPU pipelines are available. Set "auto_start": true for each of them.
intersection-cam1-gpu
intersection-cam2-gpu
intersection-cam3-gpu
intersection-cam4-gpu
Also, set "auto_start": false for the other pipelines in the same configuration file.
intersection-cam1
intersection-cam2
intersection-cam3
intersection-cam4
Start the application with:
docker compose up -d
Note: If you have multiple GPUs (integrated/discrete), please follow this DL Streamer document for selecting the GPU render device of your choice for VA codecs plugins.