The Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, state-of-the-art open model trained using the Phi-3 datasets. This dataset includes both synthetic data and filtered publicly available website data, with an emphasis on high-quality and reasoning-dense properties. The model belongs to the Phi-3 family with the Mini version in two variants 4K and 128K which is the context length (in tokens) that it can support. The model underwent a rigorous enhancement process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures. Phi-3 Mini has 3.8B parameters and is a dense decoder-only Transformer model.
The model is licensed under the MIT license.
By accessing this model, you are agreeing to the phi3 Terms and Conditions of the MIT License.
Architecture Type: Transformer
Input Format: Text
Input Parameters: None
Output Format: Text
Output Parameters: None
Supported Hardware Platform(s): RTX 4090, Ada GPUs
Supported Operating System(s): Windows
TRT-LLM Inference Engine
Windows Setup with TRT-LLM
RTX 4090