vgaraujov commited on
Commit
7e06744
1 Parent(s): 445fd01

Update Readme

Browse files
Files changed (1) hide show
  1. README.md +62 -0
README.md CHANGED
@@ -1,3 +1,65 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
5
+ - es
6
+ base_model: vgaraujov/bart-base-spanish
7
+ tags:
8
+ - generated_from_trainer
9
+ datasets:
10
+ - vgaraujov/wmt13
11
+ metrics:
12
+ - bleu
13
+ model-index:
14
+ - name: bart-base-spanish-translation-en-es
15
+ results:
16
+ - task:
17
+ name: Translation
18
+ type: translation
19
+ dataset:
20
+ name: vgaraujov/wmt13 es-en
21
+ type: vgaraujov/wmt13
22
+ config: es-en
23
+ split: validation
24
+ args: es-en
25
+ metrics:
26
+ - name: Bleu
27
+ type: bleu
28
+ value: 30.2194
29
+ widget:
30
+ - text: Hey! I am BARTO for translation.
31
  ---
32
+
33
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
34
+ should probably proofread and complete it, then remove this comment. -->
35
+
36
+ # BARTO (base-sized model) for en-es translation
37
+
38
+ This model is a fine-tuned version of [BARTO](https://huggingface.co/vgaraujov/bart-base-spanish) on a small portion of [WMT13](https://huggingface.co/datasets/vgaraujov/wmt13) es-en dataset.
39
+ It achieves the following results on the evaluation set:
40
+ - Loss: 1.7356
41
+ - Bleu: 30.2194
42
+ - Gen Len: 30.2714
43
+
44
+ ## Training procedure
45
+
46
+ ### Training hyperparameters
47
+
48
+ The following hyperparameters were used during training:
49
+ - learning_rate: 0.005
50
+ - train_batch_size: 96
51
+ - eval_batch_size: 96
52
+ - seed: 42
53
+ - gradient_accumulation_steps: 4
54
+ - total_train_batch_size: 384
55
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
56
+ - lr_scheduler_type: linear
57
+ - lr_scheduler_warmup_steps: 40000
58
+ - training_steps: 5000
59
+
60
+ ### Framework versions
61
+
62
+ - Transformers 4.33.0.dev0
63
+ - Pytorch 2.0.1+cu117
64
+ - Datasets 2.14.4
65
+ - Tokenizers 0.13.3