# FAQ 1. If you run different pipelines in a short period of time, you may encounter the following error: ![workload_error](./_images/workload_error.png)
Figure 1: Workload constraints error
This is because the maxConcurrentWorkload limitation in `AiInference.config` file. If the workloads hit the maximum, task will be canceled due to workload constrains. To solve this problem, you can kill the service with the commands below, and re-execute the command. ```bash sudo pkill Hce ``` 2. If you encounter the following error during code compilation, it is because mkl is not installed successfully: ![mkl_error](./_images/mkl_error.png)
Figure 2: Build failed due to mkl error
Run `ls /opt/intel` to check if there is a OneAPI directory in the output. If not, it means that mkl was not installed successfully. You need to reinstall mkl by following the steps below: ```bash curl -k -o GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB https://apt.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB -L sudo -E apt-key add GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB && sudo rm GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB echo "deb https://apt.repos.intel.com/oneapi all main" | sudo tee /etc/apt/sources.list.d/oneAPI.list sudo -E apt-get update -y sudo -E apt-get install -y intel-oneapi-mkl-devel lsb-release ``` 3. If the system time is incorrect, you may encounter the following errors during installation: ![oneapi_time_error](./_images/oneapi_time_error.png)
Figure 3: System Time Error
You need to set the correct system time, for example: ```bash sudo timedatectl set-ntp true ``` Then re-run the above installation command. ```bash sudo apt-get remove --purge intel-oneapi-mkl-devel sudo apt-get autoremove -y sudo apt-get install -y intel-oneapi-mkl-devel ``` 4. If you encounter the following errors during running on B580 platform: ![device_index_error](./_images/device_index_error.png)
Figure 4: Device Index Error
It may be because the iGPU is not enabled, only the B580 is enabled. You can use `lspci | grep VGA` to view the number of GPU devices on the machine. The solution is either enable iGPU in BIOS, or change the config of `Device=(STRING)GPU.1` to `Device=(STRING)GPU` in `VPLDecoderNode` and `VPLDecoderNode` in pipeline config file, for example: `ai_inference/test/configs/kitti/6C1L/localFusionPipeline.json`. 5. If you encounter the following backends mismatch errors during running pipeline: ![backends_mismatch_error](./_images/backends_mismatch_error.png)
Figure 5: Backends Mismatch Error
This is because the wrong or non-existent device is selected. We need to select the `dGPU+opencl` Backend. As shown in the figure, it should be the second device (numbered starting from 0), that is, `GPU.2`. The solution is change config `Device=(STRING)GPU.4` to `Device=(STRING)GPU.2` in `LidarSignalProcessingNode` in pipeline config file, for example: `ai_inference/test/configs/kitti/6C1L/localFusionPipeline.json`.