C/C++ Sample Apps Source Details
The DeepStream SDK package includes archives containing plugins, libraries, applications, and source code.
The sources directory is located at /opt/nvidia/deepstream/deepstream/sources
for both Debian installation (on Jetson or dGPU) and SDK Manager installation. For tar packages the source files are in the extracted deepstream package.
DeepStream Python bindings and sample applications are available as separate packages. For more information, see https://github.com/NVIDIA-AI-IOT/deepstream_python_apps.
DeepStream graphs created using the Graph Composer are listed under Reference graphs section. For more information, see the Graph Composer Introduction.
Reference test application
Path inside sources directory
Description
Sample test application 1
apps/sample_apps/deepstream-test1
Sample of how to use DeepStream elements for a single H.264 stream: filesrc → decode → nvstreammux → nvinfer or nvinferserver (primary detector) → nvdsosd → renderer. This app uses resnet18_trafficcamnet.etlt for detection.
Sample test application 2
apps/sample_apps/deepstream-test2
Sample of how to use DeepStream elements for a single H.264 stream: filesrc → decode → nvstreammux → nvinfer or nvinferserver (primary detector) → nvtracker → nvinfer or nvinferserver (secondary classifier) → nvdsosd → renderer. This app uses resnet18_trafficcamnet.etlt for detection and 2 classifier models (i.e., resnet18_vehiclemakenet.etlt, resnet18_vehicletypenet.etlt).
Sample test application 3
apps/sample_apps/deepstream-test3
Builds on deepstream-test1 (simple test application 1) to demonstrate how to:
Use multiple sources in the pipeline.
Use a uridecodebin to accept any type of input (e.g. RTSP/File), any GStreamer supported container format, and any codec.
Configure Gst-nvstreammux to generate a batch of frames and infer on it for better resource utilization.
Extract the stream metadata, which contains useful information about the frames in the batched buffer.
This app uses resnet18_trafficcamnet.etlt for detection.
Sample test application 4
apps/sample_apps/deepstream-test4
Builds on deepstream-test1 for a single H.264 stream: filesrc, decode, nvstreammux, nvinfer or nvinferserver, nvdsosd, renderer to demonstrate how to:
Use the Gst-nvmsgconv and Gst-nvmsgbroker plugins in the pipeline.
Create NVDS_META_EVENT_MSG type metadata and attach it to the buffer.
Use NVDS_META_EVENT_MSG for different types of objects, e.g. vehicle and person.
Implement “copy” and “free” functions for use if metadata is extended through the extMsg field.
This app uses resnet18_trafficcamnet.etlt for detection.
Sample test application 5
apps/sample_apps/deepstream-test5
Builds on top of deepstream-app. Demonstrates:
Use of Gst-nvmsgconv and Gst-nvmsgbroker plugins in the pipeline for multistream.
How to configure Gst-nvmsgbroker plugin from the config file as a sink plugin (for KAFKA, Azure, etc.).
How to handle the RTCP sender reports from RTSP servers or cameras and translate the Gst Buffer PTS to a UTC timestamp.
For more details refer the RTCP Sender Report callback function
test5_rtcp_sender_report_callback()
registration and usage indeepstream_test5_app_main.c
. GStreamer callback registration with rtpmanager element’s “handle-sync” signal is documented inapps-common/src/deepstream_source_bin.c
.This app uses resnet18_trafficcamnet.etlt for detection.
apps/sample_apps/deepstream-3d-lidar-sensor-fusion
The sample app deepstream-3d-lidar-sensor-fusion showcases a multi-modal sensor fusion pipeline for LiDAR and camera data using the DS3D framework. There are 2 pipelines integrated into this sample, ds3d multi-modal bevfusion with 6 cameras plus a lidar data fusion and 3D detection; v2x model Processes data from a single camera and a LiDAR, utilizing a batch size of 4.
AMQP protocol test application
libs/amqp_protocol_adaptor
Application to test AMQP protocol. This app uses resnet18_trafficcamnet.etlt for detection.
Azure MQTT test application
libs/azure_protocol_adaptor
Test application to show Azure IoT device2edge messaging and device2cloud messaging using MQTT. This app uses resnet18_trafficcamnet.etlt for detection.
DeepStream reference application
apps/sample_apps/deepstream-app
Source code for the DeepStream reference application. This app uses resnet18_trafficcamnet.etlt for detection and 2 classifier models (i.e., resnet18_vehiclemakenet.etlt, resnet18_vehicletypenet.etlt).
UFF SSD detector
sources/objectDetector_SSD
Configuration files and custom library implementation for the SSD detector model.
Yolo detector
sources/objectDetector_Yolo
Configuration files and custom library implementation for the Yolo models, currently Yolo v2, v2 tiny, v3, and v3 tiny.
Dewarper example
apps/sample_apps/deepstream-dewarper-test
Demonstrates dewarper functionality for single or multiple 360-degree camera streams. Reads camera calibration parameters from a CSV file and renders aisle and spot surfaces on the display.
Optical flow example
apps/sample_apps/deepstream-nvof-test
Demonstrates optical flow functionality for single or multiple streams. This example uses two GStreamer plugins (Gst-nvof and Gst-nvofvisual). The Gst-nvof element generates the MV (motion vector) data and attaches it as user metadata. The Gst-nvofvisual element visualizes the MV data using a predefined color wheel matrix.
Custom meta data example
apps/sample_apps/deepstream-user-metadata-test
Demonstrates how to add custom or user-specific metadata to any component of DeepStream. The test code attaches a 16-byte array filled with user data to the chosen component. The data is retrieved in another component. This app uses resnet18_trafficcamnet.etlt for detection.
MJPEG and JPEG decoder and inferencing example
apps/sample_apps/deepstream-image-decode-test
Builds on deepstream-test3 to demonstrate image decoding instead of video. This example uses a custom decode bin so the MJPEG codec can be used as input. This app uses resnet18_trafficcamnet.etlt for detection.
Image/Video segmentation example
apps/sample_apps/deepstream-segmentation-test
Demonstrates segmentation of multi-stream video or images using a semantic or industrial neural network and rendering output to a display. This app uses unet_output_graph.uff for industrial and unetres18_v4_pruned0.65_800_data.uff for semantic usecases.
Handling metadata before Gst-nvstreammux
apps/sample_apps/deepstream-gst-metadata-test
Demonstrates how to set metadata before the Gst-nvstreammux plugin in the DeepStream pipeline, and how to access it after Gst-nvstreammux. This app uses resnet18_trafficcamnet.etlt for detection.
Gst-nvinfer tensor meta flow example
apps/sample_apps/deepstream-infer-tensor-meta-app
Demonstrates how to flow and access nvinfer tensor output as metadata. NOTE: This binary is not packaged due to OpenCV deprecation. This app needs to be compiled by the user. This app uses resnet18_trafficcamnet.etlt for detection and 2 classifier models (i.e., resnet18_vehiclemakenet.etlt, resnet18_vehicletypenet.etlt).
Preprocess example
apps/sample_apps/deepstream-preprocess-test
Demonstrates inference on preprocessed ROIs configured for the streams. This app uses resnet18_trafficcamnet.etlt for detection.
3D action recognition Reference app
apps/sample_apps/deepstream-3d-action-recognition
Demonstrates a sequence batching based 3D or 2D model inference pipeline for action recognition. It also includes a sequence based preprocess custom lib for NCSHW temporal batching. Refer Prerequisites in README before running the application. This app uses resnet18_2d_rgb_hmdb5_32.etlt for 2D and resnet18_3d_rgb_hmdb5_32.etlt for 3D action recognition.
Analytics example
apps/sample_apps/deepstream-nvdsanalytics-test
Demonstrates batched analytics like ROI filtering, Line crossing, direction detection and overcrowding. This app uses resnet18_trafficcamnet.etlt for detection.
OpenCV example
apps/sample_apps/deepstream-opencv-test
Demonstrates the use of OpenCV in dsexample plugin. Need to compile dsexample with flag
WITH_OPENCV=1
. This app uses resnet18_trafficcamnet.etlt for detection.Image as Metadata example
Apps/sample_apps / deepstream-image-meta-test
Demonstrates how to attach encoded image as meta data and save the images in jpeg format. This app uses resnet18_trafficcamnet.etlt for detection.
Appsrc and Appsink example
apps/sample_apps/deepstream-appsrc-test
Demonstrates AppSrc and AppSink usage for consuming and giving data from non DeepStream code respectively. This app uses resnet18_trafficcamnet.etlt for detection.
Cuda Appsrc and Appsink example
apps/sample_apps/deepstream-appsrc-cuda-test
Demonstrates how cuda frames acquired from outside DeepStream can be fed to a DeepStream pipeline.
Transfer learning example
apps/sample_apps/ deepstream-transfer-learning-app
Demonstrates a mechanism to save the images for objects which have lesser confidence and the same can be used for training further. This app uses resnet18_trafficcamnet.etlt for detection.
Mask-RCNN example
apps/sample_apps/ deepstream-mrcnn-test
Demonstrates Instance segmentation using Mask-RCNN model. NOTE: This binary is not packaged due to OpenCV deprecation. This app needs to be compiled by the user.
DeepStream Audio Reference Application
apps/sample_apps/deepstream-audio
Source code for the DeepStream reference application demonstrating audio analytics pipeline. This app uses SONYC audio model to classify labels.
Smart Record example
apps/sample_apps/deepstream-testsr
Demonstrates event based smart record functionality. This app uses resnet18_trafficcamnet.etlt for detection.
Automatic Speech Recognition
apps/audio_apps/deepstream-asr-app
Demonstrates Automatic Speech Recognition functionality. Note: This application requires Riva ASR services to be available . Refer Prequisites in README before running the application. The default model for this application is Jasper, other options are CitriNet and QuartzNet.
Text To Speech Conversion (Alpha)
apps/audio_apps/deepstream-asr-tts-app
Demonstrates Text To Speech conversion functionality along with Automatic Speech Recognition. Note: This application requires Riva TTS and ASR services to be available. Refer prerequisites in README before running the application. This application uses the CitriNet model for ASR and FastPitch, HiFi-GAN models for TTS.
Audio+video+Text Synchronization (Alpha)
apps/sample_apps/deepstream-avsync-app
Demonstrates synchronization of audio, video and text output from nvdsasr in DeepStream pipeline. Note: This application requires Riva ASR services to be available. Refer prerequisites in README before running the application. This app uses Jasper models for Speech Recognition.
DeepStream NMOS Application
apps/sample_apps/deepstream-nmos
This application demonstrates how to create a DeepStream app as an NMOS Node. It uses a library (NvDsNmos) which provides the APIs to create, destroy and internally manage the NMOS Node. The NMOS Node can automatically discover and register with an NMOS Registry on the network using the AMWA IS-04 Registration API.
It also shows how to create various Video and Audio pipelines, run them simultaneously and reconfigure them based on NMOS events such as AMWA IS-05 Connection API requests from an NMOS Controller.
DeepStream UCX test
apps/sample_apps/deepstream-ucx-test
Demonstrates how to use the communication plugin gst-nvdsucx with DeepStream SDK. The application has been validated with kernel-5.15.
DeepStream 3D Depth Camera Reference App
apps/sample_apps/deepstream-3d-depth-camera
Demonstrates how to setup depth capture, depth render, 3D-point-cloud processing and 3D-points render pipelines over DS3D interfaces and custom-libs of ds3d::dataloader, ds3d::datafilter and ds3d::datarender. See more details in DeepStream 3D Depth Camera App
DeepStream 3D Lidar Data Inferencing Reference App
apps/sample_apps/deepstream-lidar-inference-app
Demonstrates how to read in point cloud data, inference pont cloud data with pointpillar 3D objects detection model with Triton, render point cloud data and 3D objects with GLES. The whole application is based on DS3D interfaces and custom-libs of ds3d::dataloader, ds3d::datafilter and ds3d::datarender. See more details in DeepStream 3D Lidar Inference App
Triton Onnx YOLO-v3
sources/TritonOnnxYolo
Configuration files and custom library implementation for the ONNX YOLO-V3 model. Demonstrates how to use DS-Triton to run models with dynamic-sized output tensors and how to implement custom-lib to run ONNX YoloV3 models with multi-input tensors and how to postprocess mixed-batch tensor data and attach them into nvds metadata
Deepstream Server Application
apps/sample_apps/deepstream-server
Demonstrates REST API support to control DeepStream pipeline on-the-fly.
DeepStream Can Orientation Sample App
apps/sample_apps/deepstream-can-orientation-app
Demonstrates can orientation detection with CV-based VPI template matching algorithm. VPI template matching is implemented with DeepStream video template plugin. See more details in apps/sample_apps/deepstream-can-orientation-app/README
Triton Ensemble Model Example
sources/TritonBackendEnsemble
Configuration files, Triton custom C++ backend implementation and custom library implementation for Triton ensemble model example. Demonstrates use of Triton ensemble models with gst-nvinferserver plugin and how to implement custom Triton C++ backend to access DeespStream metadata like stream ID using multi-input tensors.
deepstream-multigpu-nvlink-test
apps/sample_apps/deepstream-multigpu-nvlink-test
Uses gst-nvdsxfer plugin to simulate pipelines with NVLINK enabled multi-gpu setup to achieve better performance. User can use “position” param of nvxfer config section from dsmultigpu_config.yml file to simulate gst-nvxfer plugin supported various multi-gpu usecase pipelines.
deepstream-ipc-test-app
apps/sample_apps/deepstream-ipc-test-app
Demonstrates decoder buffer sharing IPC use-case on Jetson platform for live streams to optimize NVDEC HW utilization. This example uses IPC sink and IPC source element to interconnect GStreamer pipelines that run in different processes.
Note
Apps which write output files (example: deepstream-image-meta-test
, deepstream-testsr
, deepstream-transfer-learning-app
) should be run with sudo
permission.
Plugin and Library Source Details
The following table describes the contents of the sources directory except for the reference test applications:
Plugin or library
Path inside sources directory
Description
DsExample GStreamer plugin
gst-plugins/gst-dsexample
Template plugin for integrating custom algorithms into DeepStream SDK graph.
GStreamer Gst-nvmsgconv plugin
gst-plugins/gst-nvmsgconv
Source code for the GStreamer Gst-nvmsgconv plugin to convert metadata to schema format.
GStreamer Gst-nvmsgbroker plugin
gst-plugins/gst-nvmsgbroker
Source code for the GStreamer Gst-nvmsgbroker plugin to send data to the server.
GStreamer Gst-nvdspreprocess plugin
gst-plugins/gst-nvdspreprocess
Source code for the GStreamer Gst-nvdspreprocess plugin for preprocessing on the predefined ROIs.
GStreamer Gst-nvinfer plugin
gst-plugins/gst-nvinfer
Source code for the GStreamer Gst-nvinfer plugin for inference.
GStreamer Gst-nvinferserver plugin
gst-plugins/gst-nvinferserver
Source code for the GStreamer Gst-nvinferserver plugin for inference using Triton Inference Server.
GStreamer Gst-nvdsosd plugin
gst-plugins/gst-nvdsosd
Source code for the GStreamer Gst-nvdsosd plugin to draw bboxes, text and other objects.
Gstreamer Gst-nvdewarper plugin
gst-plugins/gst-nvdewarper
Source code for the GStreamer Gst-nvdewarper plugin to dewarp frames
NvDsInfer library
libs/nvdsinfer
Source code for the NvDsInfer library, used by the Gst-nvinfer GStreamer plugin.
NvDsInferServer library
libs/nvdsinferserver
Source code for the NvDsInferServer library, used by the Gst-nvinferserver GStreamer plugin.
NvDsNmos library
libs/nvdsnmos
Source code for the NvDsNmos library, demonstrated by the DeepStream NMOS Application.
NvMsgConv library
libs/nvmsgsconv
Source code for the NvMsgConv library, required by the Gst-nvmsgconv GStreamer plugin.
Kafka protocol adapter
libs/kafka_protocol_adapter
Protocol adapter for Kafka.
nvds_rest_server library
libs/nvds_rest_server
Source code for the rest server.
nvds_customhelper
libs/gstnvdscustomhelper
Source code for “nvdsmultiurisrcbin helper” and custom “gst-events, gst-messages and common configs” required for rest server.
nvdsinfer_customparser
libs/nvdsinfer_customparser
Custom model output parsing example for detectors and classifiers.
Gst-v4l2
See the note below 1
Source code for v4l2 codecs.
Gstreamer gst-nvdsvideotemplate plugin
gst-plugins/gst-nvdsvideotemplate
Source code for template plugin to implement video custom algorithms (non Gstreamer based)
NvDsVideoTemplate custom library
gst-plugins/gst-nvdsvideotemplate/customlib_impl
Source code for custom library to implement video custom algorithms
Gstreamer gst-nvdsaudiotemplate plugin
gst-plugins/gst-nvdsaudiotemplate
Source code for template plugin to implement audio custom algorithms (non Gstreamer based)
NvDsVideoTemplate custom library
gst-plugins/gst-nvdsaudiotemplate/customlib_impl
Source code for custom library to implement audio custom algorithms
Gstreamer gst-nvdsmetautils
gst-plugins/gst-nvdsmetautils
Source code for Gstreamer Gst-nvdsmetainsert and Gst-nvdsmetaextract plugins to process metadata
NvDsMetaUtils SEI serialization library
gst-plugins/gst-nvdsmetautils/sei_serialization
Source code for custom meta de/serialization to embed in encoded bitstream as SEI data, required by Gst-nvdsmetautils plugins
NvDsMetaUtils Audio serialization library
gst-plugins/gst-nvdsmetautils/audio_metadata_serialization
Source code for Audio NvDsFrameMeta de/serialization, required by Gst-nvdsmetautils plugins
NvDsMetaUtils Video serialization library
gst-plugins/gst-nvdsmetautils/video_metadata_serialization
Source code for Video NvDsFrameMeta & NvDsObjectMeta de/serialization, required by Gst-nvdsmetautils plugins
Gstreamer gst-nvvideotestsrc plugin
gst-plugins/gst-nvvideotestsrc
Source code to generate video test data in a variety of formats and patterns that is written directly to GPU output buffers
Gstreamer gst-nvdsspeech plugin
gst-plugins/gst-nvdsspeech
Interface for custom low level Automatic Speech Recognition (ASR) library that can be loaded by the Gst-nvdsasr plugin
Gstreamer gst-nvdstexttospeech plugin
gst-plugins/gst-nvdstexttospeech
Interface for custom low level Text To Speech (TTS) library that can be loaded by the Gst-nvds_text_to_speech plugin
Gstreamer gst-nvdspostprocess plugin
gst-plugins/gst-nvdspostprocess
Source code for the plugin and low level lib to provide a custom library interface for post processing on Tensor output of inference plugins (nvinfer/nvinferserver)
Gstreamer gst-nvtracker plugin
gst-plugins/gst-nvtracker
Source code for the plugin to track the detected objects with persistent (possibly unique) IDs over time
Gstreamer gst-nvdsanalytics plugin
gst-plugins/gst-nvdsanalytics
Interface for performing analytics on metadata attached by nvinfer (primary detector) and nvtracker
Gstreamer gst-nvstreammux New plugin
gst-plugins/gst-nvmultistream2
Source code for the plugin to form a batch of frames from multiple input sources
DS3D lidar file reader custom lib
libs/ds3d/dataloader/lidarsource
Source code for DS3D lidar file reader with an implementaion of ds3d::dataloader interface, developer could follow it to implement custom lidar device catpure, see lib details in Custom Dataloader libnvds_lidarfileread Configuration Specifications
DS3D lidar data reader custom process lib
libs/ds3d/datafilter/lidar_preprocess
Source code for DS3D lidar data processing with an implementaion of ds3d::datafilter interface
DS3D lidar detection postprocess custom lib
libs/ds3d/inference_custom_lib/ds3d_lidar_detection_postprocess
Source code for DS3D lidar and sensor fusion(bevfusion) postprocessing custom lib
DS3D lidar v2xfusion postprocess custom lib
libs/ds3d/inference_custom_lib/ds3d_v2x_infer_custom_postprocess
Source code for DS3D lidar and sensor fusion(v2xfusion) postprocessing custom lib
DS3D lidar v2xfusion preprocess custom lib
ds3d/inference_custom_lib/ds3d_v2x_infer_custom_preprocess
Source code for DS3D sensor fusion(v2xfusion) preprocess custom lib
Footnotes
- 1
Gst-v4l2 sources are not present in DeepStream package. To download, follow these steps:
In the
Search filter
field, enterL4T
Select the appropriate item for L4T Release
36.3
.Search for
L4T Driver Package (BSP) Sources
Download the file and un-tar it, to get the
.tbz2
file.Expand the
.tbz2
file.Gst-v4l2
source files are ingst-nvvideo4linux2_src.tbz2
. libnvv4l2 sources are present inv4l2_libs_src.tbz2