IProject-10 commited on
Commit
4f4c150
1 Parent(s): 0b870e9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -21,7 +21,7 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  ## Model description
23
 
24
- BERTbase fine-tuned on SQuAD 2.0 : Encoder-based Transformer Language model, pretrained with Masked Language Modeling and Next Sentence Prediction.<br>
25
  Suitable for Question-Answering tasks, predicts answer spans within the context provided.<br>
26
 
27
  **Language model:** roberta-base
@@ -40,10 +40,12 @@ For Question-Answering -
40
  from transformers import pipeline
41
  model_checkpoint = "IProject-10/roberta-base-finetuned-squad2"
42
  question_answerer = pipeline("question-answering", model=model_checkpoint)
 
43
  context = """
44
  🤗 Transformers is backed by the three most popular deep learning libraries — Jax, PyTorch and TensorFlow — with a seamless integration
45
  between them. It's straightforward to train your models with one before loading them for inference with the other.
46
  """
 
47
  question = "Which deep learning libraries back 🤗 Transformers?"
48
  question_answerer(question=question, context=context)
49
  ```
 
21
 
22
  ## Model description
23
 
24
+ RoBERTa-base fine-tuned on SQuAD 2.0 : Encoder-based Transformer Language model, pretrained with Dynamic Masking, No Next Sentence Prediction and increased Batch size compared to BERT.<br>
25
  Suitable for Question-Answering tasks, predicts answer spans within the context provided.<br>
26
 
27
  **Language model:** roberta-base
 
40
  from transformers import pipeline
41
  model_checkpoint = "IProject-10/roberta-base-finetuned-squad2"
42
  question_answerer = pipeline("question-answering", model=model_checkpoint)
43
+
44
  context = """
45
  🤗 Transformers is backed by the three most popular deep learning libraries — Jax, PyTorch and TensorFlow — with a seamless integration
46
  between them. It's straightforward to train your models with one before loading them for inference with the other.
47
  """
48
+
49
  question = "Which deep learning libraries back 🤗 Transformers?"
50
  question_answerer(question=question, context=context)
51
  ```