--- license: cc-by-nc-sa-4.0 datasets: - wi_locness - matejklemen/falko_merlin - paws - paws-x - asset language: - en - de - es - ar - ja - ko - zh metrics: - bleu - rouge - sari - accuracy library_name: transformers --- # Model Card for mEdIT-xxl This model was obtained by fine-tuning the `MBZUAI/bactrian-x-llama-13b-lora` model on the mEdIT dataset. **Paper:** mEdIT: Multilingual Text Editing via Instruction Tuning **Authors:** Vipul Raheja, Dimitris Alikaniotis, Vivek Kulkarni, Bashar Alhafni, Dhruv Kumar ## Model Details ### Model Description - **Language(s) (NLP)**: Arabic, Chinese, English, German, Japanese, Korean, Spanish - **Finetuned from model:** `MBZUAI/bactrian-x-llama-13b-lora` ### Model Sources - **Repository:** https://github.com/vipulraheja/medit - **Paper:** TBA ## How to use We release the best-performing models presented in our paper.