Linux / amd64
Linux / arm64
PyTorch is a GPU-accelerated tensor computational framework that offers a high degree of flexibility and speed for deep learning. It integrates seamlessly with popular Python libraries such as NumPy, SciPy, and Cython, extending its functionality to meet the diverse needs of users. PyTorch also employs a tape-based system for automatic differentiation at both the functional and neural network layer level, ensuring accelerated NumPy-like functionality.
Before you can run an NGC deep learning framework container, your Docker environment must support NVIDIA GPUs. To run a container, issue the appropriate command as explained in the Running A Container chapter in the NVIDIA Containers And Frameworks User Guide and specify the registry, repository, and tags. For more information about using NGC, refer to the NGC Container User Guide.
Procedure
docker pull
command.Interactive mode:
docker run --gpus all -it --rm -v local_dir:container_dir nvcr.io/nvaie/pytorch:xx.xx-py3
Non-interactive mode:
docker run --gpus all --rm -v local_dir:container_dir nvcr.io/nvaie/pytorch:xx.xx-py3
Where:
-it
means run in interactive mode
--rm
will delete the container when finished
-v
is the mounting directory
local_dir
is the directory or file from your host system (absolute path) that you want to access from inside your container. For example, the local_dir
in the following path is /home/jsmith/data/mnist
.
-v /home/jsmith/data/mnist:/data/mnist
If you are inside the container, for example, `ls /data/mnist`, you will see the same files as if you issued the `ls /home/jsmith/data/mnist` command from outside the container.
container_dir
is the target directory when you are inside your container. For example, /data/mnist
is the target directory in the example:
-v /home/jsmith/data/mnist:/data/mnist
xx.xx
is the container version. For example, 21.07
.
command
is the command you want to run in the image.
You might want to pull in data and model descriptions from locations outside the container for use by Torch. To accomplish this, the
easiest method is to mount one or more host directories as Docker data volumes. You have pulled the latest files and run the container image.
Note: DIGITS uses shared memory to share data between processes. For example, if you use Torch multiprocessing for multi-threaded data loaders, the default shared memory segment size that the container runs with may not be enough. Therefore, you should increase the shared memory size by issuing either:
--ipc=host
or
--shm-size=
in the command line to:
docker run --gpus all
/workspace/README.md
inside the container for information on customizing your PyTorch image.For the latest Release Notes, see the PyTorch Release Notes Documentation website. For a full list of the supported software and specific versions that come packaged with this framework based on the container image, see the Frameworks Support Matrix.
For more information about PyTorch, including tutorials, documentation, and examples, see:
For the optimized performance, it is highly recommended to deploy the supported NVIDIA AI Enterprise Infrastructure software in conjunction with your AI software. This release is compatible with NVIDIA AI Enterprise Infrastructure 4.4.
Please review the Security Scanning tab to view the latest security scan results. For certain open-source vulnerabilities listed in the scan results, NVIDIA provides a response in the form of a Vulnerability Exploitability eXchange (VEX) document. The VEX information can be reviewed and downloaded from the Security Scanning tab.
Collection of ftrace events may not work correctly, a newer version such as Nsight Systems 2024.5.4 from JetPack 6.1 or JetPack 5.1 could be used instead to collect ftrace events. Profiling from Nsight Systems GUI on IGX with a discrete GPU might not work, as well as connecting to such a devkit from a Ubuntu x86_64 host over SSH. In this case, please use the Nsight Systels command line (nsys) directly on the target.
By pulling and using the container, you accept the terms and conditions of this End User License Agreement.