Graphcore
Jinchen commited on
Commit
9ed4545
1 Parent(s): 19ccb6c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -10,7 +10,7 @@ Through HuggingFace Optimum, Graphcore released ready-to-use IPU-trained model c
10
 
11
 
12
  ## Model description
13
- GPT2 is a large transformer-based language model. It is built using transformer decoder blocks. BERT, on the other hand, uses transformer encoder blocks. It adds Layer normalisation to the input of each sub-block, similar to a pre-activation residual networks and additional layer normalisation.
14
 
15
  Paper link : [A ConvNet for the 2020s](https://arxiv.org/pdf/2201.03545.pdf)
16
 
 
10
 
11
 
12
  ## Model description
13
+
14
 
15
  Paper link : [A ConvNet for the 2020s](https://arxiv.org/pdf/2201.03545.pdf)
16