yjgwak commited on
Commit
35b3e2b
1 Parent(s): f2b117c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +75 -0
README.md CHANGED
@@ -1,3 +1,78 @@
1
  ---
 
 
2
  license: cc-by-sa-4.0
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+
3
+ language: ko
4
  license: cc-by-sa-4.0
5
+ tags:
6
+ - korean
7
+ - klue
8
+ - squad-kor-v1
9
+ mask_token: "[MASK]"
10
+ widget:
11
+ - text: 대한민국의 수도는 [MASK] 입니다.
12
  ---
13
+
14
+ # KLUE BERT base Finetuned on squad-kor-v1
15
+
16
+ ## Table of Contents
17
+ - [Model Details](#model-details)
18
+ - [How to Get Started With the Model](#how-to-get-started-with-the-model)
19
+ - [Uses](#uses)
20
+ - [Training](#training)
21
+ - [Evaluation](#evaluation)
22
+ - [Technical Specifications](#technical-specifications)
23
+ - [Citation Information](#citation-information)
24
+
25
+ ## Model Details
26
+
27
+ **Model Description:** This model is the KLUE BERT base, fine-tuned on the squad-kor-v1 dataset for Korean question answering tasks.
28
+
29
+ - **Developed by:** [Yeongjin Gwak](https://yjgwak.github.io/)
30
+ - **Model Type:** Transformer-based language model
31
+ - **Language(s):** Korean
32
+ - **License:** cc-by-sa-4.0
33
+ - **Parent Model:** See the [KLUE BERT base model](https://huggingface.co/klue/bert-base) for more information about the parent model.
34
+
35
+ ## How to Get Started With the Model
36
+
37
+ ```python
38
+ from transformers import AutoModel, AutoTokenizer
39
+
40
+ model = AutoModel.from_pretrained("yjgwak/klue-bert-base-finetuned-squard-kor-v1")
41
+ tokenizer = AutoTokenizer.from_pretrained("yjgwak/klue-bert-base-finetuned-squard-kor-v1")
42
+ ```
43
+
44
+ ## Uses
45
+
46
+ #### Direct Use
47
+
48
+ This model is specialized for the task of question answering in Korean. Users can employ this model to extract answers from passages or documents in Korean when provided with relevant questions.
49
+
50
+ #### Misuse and Out-of-scope Use
51
+
52
+ The model should not be used for tasks other than question answering without further fine-tuning. Using this model for generating long-form content or for tasks it wasn't fine-tuned on may result in suboptimal results.
53
+
54
+ ## Training
55
+
56
+ #### Training Data
57
+
58
+ The model was fine-tuned on the `squad-kor-v1` dataset, which is the Korean version of the popular SQuAD dataset used for question answering tasks.
59
+
60
+ #### Training Procedure
61
+
62
+ The original BERT training methodology was adopted with the difference being the dataset used for fine-tuning. The model was trained to minimize the cross-entropy loss between predicted answers and ground truth answers in the `squad-kor-v1` dataset.
63
+
64
+ ## Evaluation
65
+
66
+ [Provide details of any evaluation metrics, results, or testing data used to assess the performance of the model after fine-tuning. If this hasn't been done yet, you can mention that the evaluation is pending.]
67
+
68
+ ## Technical Specifications
69
+
70
+ See the [original KLUE BERT base model card](https://huggingface.co/klue/bert-base) for details on the underlying architecture and technical specifications.
71
+
72
+ ## Citation Information
73
+
74
+ Please cite the [original KLUE paper](https://arxiv.org/abs/2105.09680) and any other relevant resources or papers associated with the `squad-kor-v1` dataset.
75
+
76
+ ---
77
+
78
+ Once you have filled in the specifics (like the developer's name or organization and evaluation details), you can publish this model card alongside your model on the Hugging Face Model Hub.