How It Works#
This section provides a high-level view of how the application integrates with a typical system architecture.

Inputs#
Video recordings are used to simulate a live feed from cameras deployed at a toll. The application can be configured to work with live cameras.
Video Files - Tolling cameras that capture videos simultaneously from front, rear and side profiles.
Scene Database - Pre-configured intersection scene with satellite view of tolling area, calibrated cameras and regions of interest.
Core (Processing)#
Video Analytics - Deep Learning Streamer Pipeline Server (DL Streamer Pipeline Server) utilizes a pre-trained object detection model to generate object detection metadata and and a local NTP server for synchronized timestamps. This metadata is published to the MQTT broker.
Sensor Fusion - Scene Controller Microservice fuses the metadata from video analytics utilizing scene data obtained through the Scene Management API. It uses the fused tracks and the configured analytics (regions of interest) to generate events that are published to the MQTT broker.
Aggregate Scene Analytics - Region of interests analytics are read from the MQTT broker and stored in an InfluxDB bucket that enables time series analysis through Flux queries.
Live Feed Output#
Fused object tracks are available on the MQTT broker and visualized through the Scene Management UI.
Aggregated toll analytics are visualized through a Grafana dashboard.
Workflow#
Video loops or RTSP is fed into DL Streamer.
Trained AI models detect vehicles and license plates.
Metadata is published to MQTT.
Intel® SceneScape maps detections to scene regions to get exact location of objects on the scene.
Exit events are generated when vehicles leave the region.
Node-RED processes only finalized exit events by subscribing to Intel® SceneScape topics.
Data is written to InfluxDB for system to access for consistent information.
Grafana visualizes real time and historical data enabling access to metrics and vehicle details.