Model Developer: Mistralai
The Mistral-7B-Instruct-v0.3 INT4 ONNX model is the quantized version of the Mistral-7B-Instruct-v0.3 model, which is an instruct fine-tuned version of the Mistral-7B-v0.3 model used for text generation and question answering. Quantization is done with TensorRT Model Optimizer-Windows.
This model is ready for commercial use.
Steps followed to generate this quantized model:
Download Mistral-7B-Instruct-v0.3 model in Pytorch bfloat16 format from HuggingFace.
Convert PyTorch model to ONNX FP16 using onnxruntime-genai model builder.
Quantize Mistral-7B-Instruct-v0.3 ONNX FP16 model to Mistral-7B-Instruct-v0.3 ONNX INT4 AWQ model using TensorRT Model Optimizer – Windows.
This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; see link to non-NVIDIA Mistral-7B-Instruct-v0.3 Model card.
Governing Terms: Use of this model is governed by the NVIDIA Open Model License Agreement. Additional Information: Apache License, Version 2.0.
The Mistral-7B-Instruct-v0.3 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.3.
Architecture Type: Transformer
Network Architecture: Mistral-7B
Input Type: Text
Input Format: String
Input Parameters: Sequence (1D)
Other Properties Related to Input: Supports English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai
Output Type(s): Text
Output Format: String
Output Parameters: Sequence (1D)
Supported Hardware Microarchitecture Compatibility: NVIDIA Ampere and newer GPUs. 6GB or higher Video Random Access Memory (VRAM) GPUs are recommended. Higher VRAM may be required for larger context length use cases.
Supported Operating System: Windows
Model Version: v1.0
Refer to Link for the details.
Link: cnn_dailymail
Data Collection Method by dataset: [Automated]
Labeling Method by dataset: [Unknown]
Link:MMLU.
Data Collection Method by dataset: [Unknown]
Labeling Method by dataset: [Not Applicable]
Accuracy Scores: MMLU (5 shots):
With GenAI ORT->DML backend, on a desktop RTX 4090 GPU system.
“Overall accuracy” = 60.73
Test configuration:
GPU: RTX 4090, RTX3090.
Windows 11: 23H2
NVIDIA Graphics driver: R565 or higher
Inference Backend: Onnxruntime-GenAI-DirectML
(Note: Please refer to Readme.txt for the detailed instructions.)
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
Please report security vulnerabilities or NVIDIA AI Concerns here.