Convert From TensorRT to OpenVINO#

TensorRT and OpenVINO are are model optimization and inference acceleration toolkits from NVIDIA and Intel, respectively.

ToDo…

Comparison Summary

Feature

TensorRT

OpenVINO

Vendor

NVIDIA

Intel

Hardware target

NVIDIA GPUs

Intel CPUs, iGPUs, GPUs, NPUs, VPUs, FPGAs

Model input formats

ONNX, TensorFlow, PyTorch (via ONNX)

TensorFlow, ONNX, PyTorch (via ONNX)

Optimizations

Layer fusion, precision calibration, kernel tuning

Graph simplification, quantization, fusion

Output

GPU “engine” binary

OpenVINO IR model

Precision modes

FP32, FP16, INT8

FP32, FP16, INT8

Serving integration

Triton Inference Server

OpenVINO Model Server

Conversion#

ToDo…

Migration Example: Lorem Ipsum#

Quis tempor aliquip et enim aliqua. Aliquip cupidatat laboris ipsum aute eiusmod. Aliqua est adipisicing anim consequat minim nostrud aute aute pariatur dolore eiusmod qui laborum proident. Officia nisi et esse excepteur magna elit pariatur. Dolore consectetur ipsum occaecat eu. Mollit laboris ex amet commodo eiusmod mollit. Eiusmod consectetur sit laborum elit officia aute exercitation adipisicing tempor.