NGC Catalog
CLASSIC
Welcome Guest
Containers
NeMo NIM Proxy

NeMo NIM Proxy

For copy image paths and more information, please view on a desktop device.
Logo for NeMo NIM Proxy
Features
Description
Proxy service for NIM microservices
Publisher
NVIDIA
Latest Tag
25.04
Modified
April 18, 2025
Compressed Size
36.72 MB
Multinode Support
No
Multi-Arch Support
No
25.04 (Latest) Security Scan Results

Linux / amd64

Sorry, your browser does not support inline SVG.

NeMo NIM Proxy Microservice Container

NeMo NIM Proxy microservice provides a unified access point for all NVIDIA NIM (NVIDIA Inference Microservice) deployments within your Kubernetes cluster through a single OpenAI-compatible API.

You can use the NIM Proxy to interact with multiple deployed models through standardized endpoints for retrieving model lists and making inference requests to chat completions and completions APIs.

Resources

Helm Chart | User Guide

Note: Use, distribution or deployment of this microservice in production requires an NVIDIA AI Enterprise License.

Governing Terms

The software and materials are governed by the NVIDIA Software License Agreement and the Product-Specific Terms for NVIDIA AI Products.