The Nemovision-4B-Instruct model uses the latest NVIDIA recipe for distilling, pruning and quantizing to make it small enough to be performant on a broad range of RTX GPUs with the accuracy developers need. This is a model for generating responses for roleplaying, retrieval augmented generation, and function calling with vision understanding and reasoning capabilities. VRAM usage has been minimized to approximately 3.5 GB, providing fast Time to First Token. This model is ready for commercial use.
The use of this model is governed by the NVIDIA AI Foundation Models Community License Agreement
Architecture Type: Transformer
Network Architecture
Input Type(s): Image(s), Text
Input Format(s): Red, Green, Blue (RGB), String
Input Parameters: 2D, 1D
Other Properties Related to Input: The model has a maximum of 4096 input tokens.
Output Type(s): Text
Output Format(s): String
Output Parameters: 1D
Other Properties Related to Input: The model has a maximum of 4096 input tokens. Maximum output for both versions can be set apart from input.
Single Turn
<extra_id_0>System
{system prompt}
<extra_id_1>User
<image>
{prompt}
<extra_id_1>Assistant\n
<extra_id_0>System
{system prompt}
<extra_id_1>User
{prompt}
<extra_id_1>Assistant\n
Multi-image
<extra_id_0>System
{system prompt}
<extra_id_1>User
<image>
<image>
<image>
{prompt}
<extra_id_1>Assistant\n
Multi-Turn or Few-shot
<extra_id_0>System
{system prompt}
<tool> ... </tool>
<context> ... </context>
<extra_id_1>User
{prompt}
<extra_id_1>Assistant
<toolcall> ... </toolcall>
<extra_id_1>Tool
{tool response}
<extra_id_1>Assistant\n
Runtime(s): AI Inference Manager (NVAIM) Version 1.0.0
Supported Hardware Microarchitecture Compatibility: GPU supporting DirectX 11/12 and Vulkan 1.2 or higher
[Preferred/Supported] Operating System(s):
[Preferred/Supported] Operating System(s):
NV-Pretraining and NV-VILA-SFT data were used. Additionally,the following datasets were used:
Data Collection Method by dataset:
Labeling Method by dataset:
Properties:
NV-Pretraining data was collected from 5M subsampled NV-CLIP dataset. Stage 3 NV-SFT data has 3.47M unique images and 3.78M annotations on images that only have commercial license. Trained on commercial text dataset.
Data Collection Method by dataset
Labeling Method by dataset
Properties:
A collection of different benchmarks, including academic VQA benchmarks and recent benchmarks specifically proposed for language understanding and reasoning, instruction-following, and function calling LMMs.
Benchmark | VQAv2 | GQA | SQA Image | Text VQA | POPE (Popular) | MMBench-en | SEED | SEED Image | MMMU val (beam 5) |
---|---|---|---|---|---|---|---|---|---|
Accuracy | 73.92 | 53.47 | 69.81 | 57.03 | 87.13 | 59.96 | 58.89 | 66.18 | 36.8 |
Berkeley Function Calling
Benchmark | Simple | Multiple Functions | Parallel Functions | Parallel Multiple | Relevance |
---|---|---|---|---|---|
Accuracy | 85.25 | 90 | 77.5 | 76.5 | 17.08 |
Instruction Following Eval
Benchmark | Prompt Level Accuracy | Instruction Level Accuracy |
---|---|---|
Accuracy | 46.95 | 57.79 |
Test Hardware:
Supported Hardware Platform(s): L40s, A10g, A100, H100
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
Please report security vulnerabilities or NVIDIA AI Concerns here.