NGC | Catalog
CatalogContainersTAO Toolkit CV Inference Pipeline

TAO Toolkit CV Inference Pipeline

Logo for TAO Toolkit CV Inference Pipeline
NVIDIA's Train, Adapt, and Optimize (TAO) Toolkit Computer Vision Inference Pipeline allows developers to train and deploy purpose-built AI models and build AI applications using a C++ based API.
Latest Tag
April 4, 2023
Compressed Size
2.35 GB
Multinode Support
Multi-Arch Support
v0.3-ga-client (Latest) Security Scan Results

Linux / amd64

Sorry, your browser does not support inline SVG.

TAO Toolkit Computer Vision Inference Pipeline for x86

The TAO: Train, Adapt, and Optimize (TAO) Toolkit Computer Vision Inference Pipeline for x86 requires several containers:

These containers are specifically built for x86 machines with NVIDIA GPUs. Please check the Requirements and Installation before usage.

The NVIDIA Triton Inference Server is hosted on the NVIDIA GPU Cloud (NGC). This exists as another process and serves inferences to the Client container, which houses applications and sample usage for the TAO Toolkit Computer Vision API.

The Server Utilities container contains the folder structure and libraries necessary for the NVIDIA Triton Inference Server to serve inferences. This container also allows for simple TAO model conversion from the TAO Toolkit Computer Vision Quick Start.

These containers and tags will automatically be pulled by the Quick Start Scripts without intervention from the user. Please check the guidelines provided by the Quick Start Scripts for more details.

Client Container

The Client Container provides an environment with the TAO Toolkit CV Inference Pipeline libraries and open-source demos that enable developers to build and deploy custom applications.

The included demos are as follows:

  • Body Pose Estimation
  • Emotion Classification
  • Face Detection
  • Facial Landmark Estimation
  • Gaze Estimation
  • Gesture Recognition

Each of these demos leverages appicable TAO Toolkit models, which can be retrained. These open source demos highlight the C++ API that allows for inference requests (eg. getBodyPose(), getGaze()).

These can run using a webcam device or a video file/stream. There is also an API that allows for custom image decode. More information is provided in the demo source and the API Documentation in the Quick Start. Further configuration documentation is also provided through the TAO Toolkit Documentation.

Some examples of applications that one can build are event-based applications. If a person is providing a thumbs-up gesture or is looking directly at the camera, the gesture recognition and gaze estimation APIs can allow for the detection of these events.


License for TAO Toolkit Computer Vision Inference Pipeline containers is included within the containers at workspace/TAO-CV-Inference-Pipeline-EULA.pdf. License for the pre-trained models are available with the model files. By pulling and using the TAO Toolkit Computer Vision Inference Pipeline and downloading models, you accept the terms and conditions of these licenses.

Ethical AI

NVIDIA's platforms and application frameworks enable developers to build a wide array of AI applications. Consider potential algorithmic bias when choosing or creating the models being deployed. Work with the model's developer to ensure that it meets the requirements for the relevant industry and use case; that the necessary instruction and documentation are provided to understand error rates, confidence intervals, and results; and that the model is being used under the conditions and in the manner intended.