Linux / arm64
This container is deprecated. It was released as part of the Clara Holoscan SDK v0.1 and will no longer be compatible with Clara Holoscan SDK v0.2.
This release is PyTorch built only with support for Clara AGX Hardware.
PyTorch is a GPU accelerated tensor computational framework with a Python front end. Functionality can be easily extended with common Python libraries such as NumPy, SciPy, and Cython. Automatic differentiation is done with a tape-based system at both a functional and neural network layer level. This functionality brings a high level of flexibility and speed as a deep learning framework and provides accelerated NumPy-like functionality.
Before you can run an NGC deep learning framework container, your Docker environment must support NVIDIA GPUs. To run a container, issue the appropriate command as explained in the Running A Container chapter in the NVIDIA Containers And Frameworks User Guide and specify the registry, repository, and tags. For more information about using NGC, refer to the NGC Container User Guide.
The method implemented in your system depends on the AGX OS version installed, the specific NGC Cloud Image provided by a Cloud Service Provider, or the software that you have installed in preparation for running NGC containers.
Procedure
Select the Tags tab and locate the container image release that you want to run.
In the Pull Tag column, click the icon to copy the docker pull command.
Open a command prompt and paste the pull command. The pulling of the container image begins. Ensure the pull completes successfully before proceeding to the next step.
Run the container image. To run the container, choose interactive mode or non-interactive mode.
A typical command to launch the container is:
docker run --gpus all -it --rm -v local_dir:container_dir nvcr.io/nvidia/pytorch:xx.xx-py3
A typical command to launch the container is:
docker run --gpus all -it --rm -v local_dir:container_dir nvcr.io/ea-clara-agx/agx-pytorch:xx.xx-py3
Where:
-it
means run in interactive mode
--rm
will delete the container when finished
-v
is the mounting directory
local_dir
is the directory or file from your host system (absolute path) that you want to access from inside your container. For example, the local_dir
in the following path is /home/jsmith/data/mnist
.
-v /home/jsmith/data/mnist:/data/mnist
If you are inside the container, for example, ls /data/mnist
, you will see the same files as if you issued the ls /home/jsmith/data/mnist
command from outside the container.
container_dir
is the target directory when you are inside your container. For example, /data/mnist
is the target directory in the example:
-v /home/jsmith/data/mnist:/data/mnist
xx.xx
is the container version. For example, 20.01
.
You might want to pull in data and model descriptions from locations outside the container for use by Torch. To accomplish this, the easiest method is to mount one or more host directories as Docker data volumes. You have pulled the latest files and run the container image.
Note: DIGITS uses shared memory to share data between processes. For example, if you use Torch multiprocessing for multi-threaded data loaders, the default shared memory segment size that the container runs with may not be enough. Therefore, you should increase the shared memory size by issuing either:
--ipc=host
or
--shm-size=
in the command line to:
docker run --gpus all
For the latest Release Notes, see the PyTorch Release Notes Documentation website.
For more information about PyTorch, including tutorials, documentation, and examples, see:
An End User License Agreement is included with this product. By pulling and using the containers from NGC, you accept the terms and conditions of this license.