pratikdoshi commited on
Commit
05bddc6
1 Parent(s): c66ffd6

updated hyperparam formatting

Browse files
Files changed (1) hide show
  1. README.md +12 -11
README.md CHANGED
@@ -28,17 +28,18 @@ It is trained in a manner described by [Phil Schmid here](https://www.philschmid
28
 
29
  ### Training hyperparameters
30
 
31
- The following hyperparameters were used during training:
32
- - learning_rate: 0.0002
33
- - train_batch_size: 50
34
- - eval_batch_size: 8
35
- - seed: 42
36
- - gradient_accumulation_steps: 2
37
- - total_train_batch_size: 100
38
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
- - lr_scheduler_type: constant
40
- - lr_scheduler_warmup_ratio: 0.03
41
- - num_epochs: 3
 
42
 
43
  ### Training results
44
 
 
28
 
29
  ### Training hyperparameters
30
 
31
+ | Hyperparameter | value |
32
+ | --- | --- |
33
+ | learning_rate | 0.0002 |
34
+ | train_batch_size | 50 |
35
+ | eval_batch_size | 8 |
36
+ | seed | 42 |
37
+ | gradient_accumulation_steps | 2 |
38
+ | total_train_batch_size | 100 |
39
+ | optimizer | Adam with betas=(0.9,0.999) and epsilon=1e-08 |
40
+ | lr_scheduler_type | constant |
41
+ | lr_scheduler_warmup_ratio | 0.03 |
42
+ | num_epochs | 3 |
43
 
44
  ### Training results
45