Model Depot
Collection
Leading generative models packaged in OpenVino format optimized for use on AI PCs
•
50 items
•
Updated
•
5
tiny-llama-chat-ov is an OpenVino int4 quantized version of TinyLlama-Chat, providing a very fast, very small inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
tiny-llama-chat is the official chat finetuned version of tiny-llama.
Base model
TinyLlama/TinyLlama-1.1B-Chat-v1.0