Filipino Sentence RoBERTa
We finetuned RoBERTa Tagalog Base (finetuned on COHFIE) on NewsPH-NLI to learn to encode filipino/tagalog sentences to sentence embeddings. We used sentence-transformers to finetune the model. All model details, training setups, and corpus details can be found in this paper: Automatic WordNet Construction using Word Sense Induction through Sentence Embeddings.
Intended uses & limitations
The intended use of this model is to extract sentence embeddings which will be used for clustering. This model may not be safe for use in production since we did not examine it for biases. Please use it with caution.
How to use
Using this model is easier when you have sentence-transformers installed:
pip install -U sentence-transformers
Here is how to use this model to encode sentences to sentence embeddings using SentenceTransformer
:
from sentence_transformers import SentenceTransformer
model = SentenceTransformer("danjohnvelasco/filipino-sentence-roberta-v1")
sentence_list = ["sentence 1", "sentence 2", "sentence 3"]
sentence_embeddings = model.encode(sentence_list)
print(sentence_embeddings)
BibTeX entry and citation info
If you use this model, please cite our work:
@misc{https://doi.org/10.48550/arxiv.2204.03251,
doi = {10.48550/ARXIV.2204.03251},
url = {https://arxiv.org/abs/2204.03251},
author = {Velasco, Dan John and Alba, Axel and Pelagio, Trisha Gail and Ramirez, Bryce Anthony and Cruz, Jan Christian Blaise and Cheng, Charibeth},
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Automatic WordNet Construction using Word Sense Induction through Sentence Embeddings},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
- Downloads last month
- 9