--- library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - ExLlamaV2 - 4bit - Mistral - Mistral-7B - quantized - exl2 - 5.0-bpw --- # Mistral-7B-Instruct-v0.2-5.0-bpw-exl2 This repo contains 4-bit quantized (using ExLlamaV2) model Mistral AI_'s Mistral-7B-Instruct-v0.2 ## Model Details - Model creator: [Mistral AI_](https://huggingface.co/mistralai) - Original model: [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) ### About 4 bit quantization using ExLlamaV2 - ExLlamaV2 github repo: [ExLlamaV2 github repo](https://github.com/turboderp/exllamav2) # How to Get Started with the Model Use the code below to get started with the model. ## How to run from Python code #### First install the package # Import ```python import torch import os from torch import bfloat16 from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline, BitsAndBytesConfig, LlamaForCausalLM ``` # Use a pipeline as a high-level helper ## Uses ### Direct Use [More Information Needed] ### Out-of-Scope Use [More Information Needed] ## Bias, Risks, and Limitations [More Information Needed] ## Evaluation #### Metrics [More Information Needed] ### Results [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]