Linux / amd64
NVIDIA NIMâ„¢, part of NVIDIA AI Enterprise, is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI model inferencing across clouds, data centers and workstations. Supporting a wide range of AI models, including open-source and NVIDIA AI Foundation and custom models, it ensures seamless, scalable AI inferencing, on-premises or in the cloud, leveraging industry standard APIs.
Audio2Face-2D, also known as Speech Live Portrait, animates a person's portrait photo using a driving audio. It also accepts tuning values to control eye blinks, head movement, and gaze of the portrait animation. This NIM offers two modes: performance and quality.
NVIDIA NIM offers prebuilt containers for AI models across computer vision, audio, LLMs, and more. Each NIM consists of a container and a model and uses a CUDA-accelerated runtime for all NVIDIA GPUs, with special optimizations available for many configurations. Whether on-premises or in the cloud, NIM is the fastest way to achieve accelerated inference at scale.
Deploying and integrating NVIDIA NIM is straightforward thanks to our industry-standard APIs. Visit the Maxine Audio2Face-2D NIM page for release documentation, deployment guides, and more.
Get access to knowledge base articles and support cases or submit a ticket.
The NIM container is governed by the NVIDIA AI Enterprise Software License Agreement; and the use of this model is governed by the NVIDIA AI Foundation Models Community License.
You are responsible for ensuring that your use of NVIDIA AI Foundation Models complies with all applicable laws.