Linux / arm64
With the DeepStream 6.3 Release, we have introduced a new NGC DeepStream SDK Collection.
This collection serves as a hub for all DeepStream assets. Make sure you check it out!
Please refer to the section below which describes the different container options offered for Jetson-based devices:
Container Name |
Architecture |
License Type |
Replaces (previous releases) |
Notes |
---|---|---|---|---|
deepstream:6.3-triton-multiarch |
Multi-Arch X86 + Jetson |
Deployment |
6.2-triton |
The DeepStream Triton container enables inference using Triton Inference Server. With Triton developers can run inference natively using TensorFlow, TensorFlow-TensorRT, PyTorch and ONNX-RT. Inference with Triton is supported in the reference application (deepstream-app) |
deepstream-l4t:6.3-samples |
Jetson |
Deployment |
N/A |
The DeepStream samples container extends the base container to also include sample applications that are included in the DeepStream SDK along with associated config files, models, and streams. This container is ideal to understand and explore the DeepStream SDK using the provided samples. |
Notes:
· Dockers from previous CUDA releases DeepStream dockers or dockers derived from previous releases (before DeepStream 6.1) will need to update their CUDA GPG key to perform software updates. You can find additional details here for details.
· These containers leverage the NVIDIA Container Runtime on Jetson, which is available for install as part of NVIDIA JetPack version 5.1.2 . The platform specific libraries and select device nodes for a particular device are mounted by the NVIDIA Container Runtime into the DeepStream container from the underlying host, thereby providing necessary dependencies (BSP Libraries) for DeepStream applications to execute within the container.
· Since Jetpack 5.1.2, NVIDIA Container Runtime no longer mounts user level libraries like CUDA, cuDNN and TensorRT from the host. These will instead be installed inside the containers.
Ensure these prerequisites are installed in your system before proceeding to the next step:
Component |
Details |
---|---|
A Jetson device running L4T BSP r35.4.1 . |
|
Codecs script |
DeepStream dockers no longer package libraries for certain multimedia operations such as: audio data parsing, CPU decode, and CPU encode. This translates into limited functionality with MP4 files. We provide a script to install these components. Make sure to execute the script within the container: |
1. From the top-right corner of this page, select the pull-down Get Container and copy the URL to the default container. Alternatively, click on View all tags to select a different container.
2. Open a command prompt on your Linux compatible system and run the following command. Ensure the pull completes successfully before proceeding to the next step.
docker pull nvcr.io/nvidia/deepstream:6.3-triton-multiarch
3. Allow external applications to connect to the host's X display:
xhost +
4. Run the docker
container (use the desired container tag in the command line below):
If using docker (recommended):
docker run -it --rm --net=host --runtime nvidia -e DISPLAY=$DISPLAY -w /opt/nvidia/deepstream/deepstream-6.3 -v /tmp/.X11-unix/:/tmp/.X11-unix nvcr.io/nvidia/deepstream:6.3-triton-multiarch
Docker command line options explained:
Option |
Description |
---|---|
-it |
means run in interactive mode |
--runtime nvidia |
Use an alternative runtime (will use the NVIDIA container runtime) |
--rm |
will delete the container when finished |
--privileged |
grants access to the container to the host resources. This flag is need to run Graph Composer from the -devel container |
-v |
Specifies the mounting directory, and used to mount host's X11 display in the container filesystem to render output videos |
--cap-add SYSLOG |
This option needs to be included to enable usage of the nvds_logger functionality inside the container |
-p |
to enable RTSP out, network port needs to be mapped from container to host to enable incoming connections using the-poption in the command line; eg:-p 8554:8554 |
See /opt/nvidia/deepstream/deepstream-6.3/README inside the container for deepstream-app usage information.
Additional argument to add to above docker command for accessing CSI Camera from Docker: -v /tmp/argus_socket:/tmp/argus_socket For USB Camera additional argument --device /dev/video
NOTE
Please refer
to /opt/nvidia/deepstream/deepstream-6.3/README
inside the container for
details on deepstream-app usage.
Limitations
· The DeepStream containers for Jetson are intended to be a deployment vehicle. Please refer “Docker Containers” Section within the DeepStream 6.3 Plugin Guide Section for instructions on how to build custom containers based on DeepStream from either Jetson device or your workstation.
· AMQP support is not included inside the container. Please refer “AMQP Protocol Adapter” Section within the DeepStream 6.3 Plugin Guide Section for instructions on how to install necessary dependencies for enabling AMQP if required.
· There are known bugs and limitations in the SDK. To learn more about those, refer to the release notes.
The following licenses apply to the DeepStream SDK assets:
Asset |
Applicable EULA |
Notes |
---|---|---|
SDK |
A copy of the license is available on the following folder of the SDK: |
|
Containers |
License grants redistribution rights allowing developers to build applications on top of the DeepStream containers |
|
Development Containers |
A development-only license. Does not allow redistribution of the container |
NOTE: By pulling, downloading, or using the DeepStream SDK, you accept the terms and conditions of the EULA licenses listed above.
Please note that all container images come with the following packages installed:
The software listed below is provided under the terms of GPLv3.
To obtain source code for software provided under licenses that require redistribution of source code, including the GNU General Public License (GPL) and GNU Lesser General Public License (LGPL), contact oss-requests@nvidia.com. This offer is valid for a period of three (3) years from the date of the distribution of this product by NVIDIA CORPORATION.
Component |
License |
---|---|
autoconf |
GPL 3.0 |
libtool |
GPL 3.0 |
libglvnd-dev |
GPL 3.0 |
libgl1-mesa-dev |
GPL 3.0 |
libegl1-mesa-dev |
GPL 3.0 |
libgles2-mesa-dev |
GPL 3.0 |
NVIDIA’s platforms and application frameworks enable developers to build a wide array of AI applications. Consider potential algorithmic bias when choosing or creating the models being deployed. Work with the model’s developer to ensure that it meets the requirements for the relevant industry and use case; that the necessary instruction and documentation are provided to understand error rates, confidence intervals, and results; and that the model is being used under the conditions and in the manner intended.