Elements#
Links under the GStreamer element name (first column of the table) contain the description of element properties, in the format generated by gst-inspect-1.0 utility.
Inference plugins#
Element |
Description |
|---|---|
Performs object detection and optionally object classification/segmentation/pose estimation. Inputs: ROIs (regions of interest) or full frame. Output: object bounding box detection along with prediction metadata. The |
|
Performs object classification/segmentation/pose estimation. Inputs: ROIs or full frame. Output: prediction metadata. The |
|
Executes any inference model and outputs raw results. Does not interpret data and does not generate metadata. The |
|
Tracks objects across video frames using zero-term or short-term tracking algorithms. Zero-term tracking assigns unique object IDs and requires object detection to run on every frame. Short-term tracking allows for tracking objects between frames, reducing the need to run object detection on each frame. |
|
Legacy plugin. Performs audio event detection using the |
|
ASR plugin. Performs audio transcription using |
|
Performs inference using GenAI models. It can be used to generate text descriptions from images or video. |
3D plugins#
Element |
Description |
|---|---|
Processes millimeter-wave (mmWave) radar signal data. Performs data reordering, pre-processing, DC (Direct Current) removal, and interfaces with the radar library to generate point clouds, clusters, and tracking data. Attaches custom metadata containing detected reflection points, clustered objects, and tracked targets to each buffer. |
|
Parses 3D LiDAR binary frames and attaches custom metadata with point cloud data. It reads raw LiDAR frames (BIN/PCD), applies stride/frame-rate thinning, and outputs buffers enriched with LidarMeta (points, frame_id, timestamps, stream_id) for downstream fusion, analytics, or visualization. |
Auxiliary plugins#
Element |
Description |
|---|---|
Adds user-defined regions of interest to perform inference on (instead of full frame). Example: monitoring road traffic in a city camera feed; splitting large image into smaller pieces, and running inference on each piece (healthcare cell analytics). |
|
Measures frames per second across multiple video streams in a single GStreamer process. |
|
Throttles the framerate of video streams by enforcing a maximum frames-per-second (FPS) rate. Useful for rate limiting in pipelines or for testing at specific processing framerates. |
|
Aggregates inference results from multiple pipeline branches. |
|
Converts the metadata structure to JSON or raw text formats. Can write output to a file. |
|
Publishes the JSON metadata to MQTT or Kafka message brokers or files. |
|
Provides a callback to execute user-defined Python functions on every frame. It is used to augment DL Streamer with user-defined algorithms (e.g. metadata conversion, inference post-processing). |
|
Provides integration with Intel RealSense cameras, enabling video and depth stream capture for use in GStreamer pipelines. |
|
Overlays the metadata on the video frame to visualize the inference results. |
|
Performs lightweight motion detection on NV12 frames and emits motion ROIs as analytics metadata. Uses VA-API acceleration when VAMemory caps are negotiated, otherwise system-memory path. |