We have introduced a new NGC DeepStream SDK Collection.
This collection serves as a hub for all DeepStream assets. Make sure you check it out!
Please refer to the section below which describes the different container options offered for NVIDIA Data Center GPUs running on x86 platform.
Container Name | Architecture | License Type | Notes |
---|---|---|---|
deepstream:8.0-triton-multiarch | Multi-Arch x86 + Jetson |
Deployment | The DeepStream Triton container enables inference using Triton Inference Server. With Triton developers can run inference natively using PyTorch and ONNX-RT. Inference with Triton is supported in the reference application (deepstream-app) |
deepstream:8.0-samples-multiarch | Multi-Arch x86 + Jetson |
Deployment | The DeepStream samples container extends the base container to also include sample applications that are included in the DeepStream SDK along with associated config files, models, and streams. This container is ideal to understand and explore the DeepStream SDK using the provided samples. |
deepstream:8.0-triton-arm-sbsa | SBSA | Deployment | SBSA container for ARM based servers and systems. |
deepstream:8.0-gc-triton-devel | x86 | Development | The DeepStream development container is the recommended container to get you started as it includes Graph Composer, the build toolchains, development libraries and packages necessary for building DeepStream reference applications within the container. This container is slightly larger in size by virtue of including the build dependencies. |
Read the DeepStream 8.0 Release Notes for full list of features of the release
NOTE: Dockers from previous CUDA releases DeepStream dockers or dockers derived from previous releases (before DeepStream 6.1) will need to update their CUDA GPG key to perform software updates. You can find additional details here for details.
Ensure these prerequisites are installed in your system before proceeding to the next step:
Component | Details |
---|---|
nvidia-docker | We recommend using Docker 20.10.13 along with the latest nvidia-container-toolkit as described in the installation steps. Usage of nvidia-docker2 packages in conjunction with prior docker versions is now deprecated. |
NVIDIA GPU Driver | Use version: 570.133.20 for production deployments for Data Center GPUs |
Codecs script | DeepStream dockers no longer package libraries for certain multimedia operations such as audio data parsing, CPU decode, and CPU encode. This translates into limited functionality with MP4 files. We provide a script to install these components. Make sure to execute
the script within the container: |
From the top-right corner of this page, select the pull-down Get Container and copy the URL to the default container. Alternatively, click on View all tags to select a different container.
Open a command prompt on your Linux compatible system and run the following command. Ensure the pull completes successfully before proceeding to the next step.
docker pull nvcr.io/nvidia/deepstream:8.0-triton-multiarch
xhost +
For x86 Systems:
docker run --gpus all -it --rm --runtime=nvidia --network=host --privileged -v
/tmp/.X11-unix:/tmp/.X11-unix -e DISPLAY=$DISPLAY -w
/opt/nvidia/deepstream/deepstream-8.0 nvcr.io/nvidia/
deepstream:8.0-triton-multiarch
For ARM SBSA Systems:
docker run -it --rm --runtime=nvidia --network=host -e
NVIDIA\_DRIVER\_CAPABILITIES=compute,utility,video,graphics --gpus all
--privileged -e DISPLAY=$DISPLAY -v /tmp/.X11-unix:/tmp/.X11-unix -v
/etc/X11:/etc/X11 nvcr.io/nvidia/deepstream:8.0-triton-arm-sbsa
Docker command line options explained:
Option | Description |
---|---|
-it | means run in interactive mode |
--gpus | This option makes GPUs accessible inside the container. It is also possible to specify a device (i.e. '"'device=0'") |
--rm | will delete the container when finished |
--privileged | grants access to the container to the host resources. This flag is needed to run Graph Composer from the -devel container |
-v | Specifies the mounting directory and it can be used to mount host's X11 display in the container filesystem to render output videos. Users can mount additional directories (using -v option) as required to easily access configuration files, models, and other resources. (i.e., use -v /home:/home) to mount the home directory into the container filesystem. |
--cap-add SYSLOG | This option needs to be included to enable usage of the nvds_logger functionality inside the container |
-p | to enable RTSP out, network port needs to be mapped from container to host to enable incoming connections using the -p option in the command line; eg :-p 8554:8554 |
NOTES:
Please refer to /opt/nvidia/deepstream/deepstream-8.0/README inside
the container for details on deepstream-app usage.
Using the Triton docker as a base image:
For creating a base image using the Triton (x86) docker as a baseline,
one approach is to use an entry point with a combined script so end
users can run a specific script for their application.
ENTRYPOINT ["/bin/sh", "-c" , "/opt/nvidia/deepstream/deepstream- 8.0/entrypoint.sh && \custom command\"]
apt-get install --reinstall libflac8 libmp3lame0 libxvidcore4 ffmpeg
DeepStream 8.0 Triton containers (x86 and SBSA) have the following CVE(s):
CVE | Description |
---|---|
CVE-2025-3887 | GStreamer H265 Codec Parsing Stack-based Buffer Overflow Remote Code Execution Vulnerability. This vulnerability allows remote attackers to execute arbitrary code on affected installations of GStreamer. This impacts Gstreamer H265 parser on Ubuntu 24.04. There is no publicly available fix for Ubuntu 24.04. Users can opt for Ubuntu Pro which has this fix included from Canonical. |
The following licenses apply to the DeepStream SDK assets:
Asset | Applicable EULA | Notes |
---|---|---|
SDK | DeepStream SDK EULA | A copy of the license is available on the following folder of the
SDK: /opt/nvidia/deepstream/deepstream-8.0/LicenseAgreement.pdf |
Containers | DeepStream NGC License | License grants redistribution rights allowing developers to build applications on top of the DeepStream containers |
Development Containers | DeepStream NGC Development License | A development-only license. Does not allow redistribution of the container |
TAO Models | NVIDIA AI Product License | All TAO pre-trained models included in the DeepStream SDK are covered by the NVIDIA AI Product License. |
NOTE: By pulling, downloading, or using the DeepStream SDK, you accept
the terms and conditions of the EULA licenses listed above.
This project will download and install additional third-party open
source software projects. Review the license terms of these open source
projects before use.
Please note that all container images come with the following packages installed:
autoconf ( license and license exception )
In addition, the (deepstream:8.0-gc-triton-devel) container includes the Vulkan Validation Layers (v1.1.123) to support the NVIDIA Graph Composer.
The software listed below is provided under the terms of GPLv3.
To obtain source code for software provided under licenses that require redistribution of source code, including the GNU General Public License (GPL) and GNU Lesser General Public License (LGPL), available under DeepStream Resources 8.0_3rd_party_oss. This offer is valid for a period of three (3) years from the date of the distribution of this product by NVIDIA CORPORATION.
Component | License |
---|---|
autoconf | GPL 3.0 |
libtool | GPL 3.0 |
libglvnd-dev | GPL 3.0 |
libgl1-mesa-dev | GPL 3.0 |
libegl1-mesa-dev | GPL 3.0 |
libgles2-mesa-dev | GPL 3.0 |
NVIDIA’s platforms and application frameworks enable developers to build a wide array of AI applications. Consider potential algorithmic bias when choosing or creating the models being deployed. Work with the model’s developer to ensure that it meets the requirements for the relevant industry and use case; that the necessary instruction and documentation are provided to understand error rates, confidence intervals, and results; and that the model is being used under the conditions and in the manner intended.