Model Depot - ONNX
Collection
Leading Models packaged in ONNX format optimized for use with AI PCs
•
20 items
•
Updated
phi-3-onnx is an ONNX int4 quantized version of Microsoft Phi-3-mini-4k-instruct, providing a very fast, very small inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
Base model
microsoft/Phi-3-mini-4k-instruct