--- language: - multilingual - en - zh license: apache-2.0 pipeline_tag: fill-mask tags: - medical --- # KBioXLM The aligned corpus constructed using the knowledge-anchored method is combined with a multi task training strategy to continue training XLM-R, thus obtaining KBioXLM. It is the first multilingual biomedical pre-trained language model we know that has cross-lingual understanding capabilities in medical domain. It was introduced in the paper [KBioXLM: A Knowledge-anchored Biomedical Multilingual Pretrained Language Model](http://arxiv.org/abs/2311.11564). and first released in [this repository](https://github.com/ngwlh-gl/KBioXLM/tree/main) ## Model description KBioXLM model can be fintuned on downstream tasks. The downstream tasks here refer to biomedical cross-lingual understanding tasks, such as biomedical entity recognition, biomedical relationship extraction and biomedical text classification. ## Usage You can follow the prompts below to load our model parameters: ```python from transformers import RobertaModel model=RobertaModel.from_pretrained('ngwlh/KBioXLM') ``` ### BibTeX entry and citation info Coming soon.