We have introduced PipeTuner Collection. PipeTuner is an automatic tuning tool that efficiently explores the parameter space and finds the optimal parameters for the pipelines, which yields the highest KPI on the dataset provided by the user. This resource page contains the documentation, sample data and config files to run PipeTuner. Visit "File Browser" to download them.
Users need to follow all the steps in this section to start tuning.
PipeTuner requires the following components on an x86_64 system:
Users need to follow below steps to sign in to an NGC account and get an API key.
$ docker login nvcr.io
Username: "$oauthtoken"
Password: "YOUR_NGC_API_KEY"
$ ngc config set
Enter API key: "YOUR_NGC_API_KEY"
Enter org: nfgnkvuikvjm
Enter team: mdx-v2-0
The sample data consists of a mini-synthetic dataset with eight 1-minute streams and config files for tuning. You can download the sample files pipe-tuner-sample.zip by clicking “Download” from this page.
Once you download the sample file, unzip the file and run setup.sh to finish sample data for either DeepStream or Metropolis Microservices.
$ unzip pipe-tuner-sample.zip
$ cd pipe-tuner-sample/scripts
$ # DeepStream or Metropolis Microservices users should run only one of the following two commands based on their usage
$ bash setup.sh deepstream # DeepStream users
$ bash setup.sh metropolis # Metropolis Microservices users
DeepStream users should see docker images like below.
$ docker images # bash setup.sh deepstream
REPOSITORY TAG
nvcr.io/nvidia/pipetuner 1.0
nvcr.io/nvidia/deepstream 7.0-triton-multiarch
Also, model files should be under the ‘models’ folder. They will be mapped into DeepStream containers during tuning.
$ ls ../models
labels.txt resnet34_peoplenet_int8.etlt resnet34_peoplenet_int8.txt resnet50_market1501_aicity156.onnx
Metropolis users should see docker images like below. The ‘models’ folder is empty because default models in mdx-perception container will be used.
$ docker images # bash setup.sh metropolis
REPOSITORY TAG
nvcr.io/nvidia/pipetuner 1.0
nvcr.io/nfgnkvuikvjm/mdx-v2-0/mdx-perception 2.1
The final directory under pipe-tuner-sample is like:
pipe-tuner-sample
├── configs
│ ├── config_CameraMatrix
│ ├── config_GuiTool
│ ├── config_MTMC
│ ├── config_PGIE
│ ├── config_PipeTuner
│ └── config_Tracker
├── data
│ ├── SDG_1min_utils
│ └── SDG_1min_videos
├── models
├── ngc_download
├── multi-camera-tracking (only for Metropolis Microservice)
└── scripts
Asset | Applicable EULA | Notes |
---|---|---|
PipeTuner Container | NVIDIA_PipeTuner_EULA | A copy of the license is available in the following path inside the container: /pipe-tuner/NVIDIA_PipeTuner_EULA.pdf |
NOTE: By pulling, downloading, or using PipeTuner, you accept the terms and conditions of the EULA licenses listed above.
For DeepStream SDK and Metropolis Microservices, please refer to their own licenses.
NVIDIA’s platforms and application frameworks enable developers to build a wide array of AI applications. Consider potential algorithmic bias when choosing or creating the models being deployed. Work with the model’s developer to ensure that it meets the requirements for the relevant industry and use case; that the necessary instruction and documentation are provided to understand error rates, confidence intervals, and results; and that the model is being used under the conditions and in the manner intended.