--- inference: false language: - bg license: mit tags: - torch --- # LLaMA-7B This repo contains a low-rank adapter for LLaMA-7b trained on Bulgarian dataset. This model was introduced in [this paper](https://arxiv.org/abs/2106.09685). ## Model description The training data is private Bulgarian dataset. ## Intended uses & limitations This is an instruction-based model similart to ChatGPT but in Bulgarian. ### How to use Here is how to use this model in PyTorch: ```bash >>> git clone https://github.com/tloen/alpaca-lora.git >>> cd alpaca-lora >>> pip install -r requirements.txt >>> >>> python generate.py \ --load_8bit \ --base_model 'yahma/llama-7b-hf' \ --lora_weights 'rmihaylov/alpaca-lora-bg-7b' \ --share_gradio ``` This will download both a base model and an adapter from huggingface. Then it will run a gradio interface for chatting. Example using this model: [Colab](https://colab.research.google.com/drive/1IPz8QqOa5ZUBz7ZyXE4hhh7XwMEH-D9S#scrollTo=onLKXdA97gjs) ### Interface ![interface](https://huggingface.co/rmihaylov/alpaca-lora-bg-7b/blob/main/example.png)