pstroe commited on
Commit
9013599
1 Parent(s): 66bffc8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -4,7 +4,7 @@ This is a Latin RoBERTa-based LM model, version 3.
4
 
5
  The intention of the Transformer-based LM is twofold: on the one hand, it will be used for the evaluation of HTR results; on the other, it should be used as a decoder for the TrOCR architecture.
6
 
7
- The training data differs from the one used in the RoBERTa Bas Latin Cased V1 and V2, and therefore also by what is used by [Bamman and Burns (2020)](https://arxiv.org/pdf/2009.10053.pdf). We exclusively used the text from the [Corpus Corporum](https://www.mlat.uzh.ch).
8
 
9
  The overall corpus contains 1.5G of text data (3x as much as has been used for V2 and very likely of better quality).
10
 
 
4
 
5
  The intention of the Transformer-based LM is twofold: on the one hand, it will be used for the evaluation of HTR results; on the other, it should be used as a decoder for the TrOCR architecture.
6
 
7
+ The training data differs from the one used in the RoBERTa Bas Latin Cased V1 and V2, and therefore also by what is used by [Bamman and Burns (2020)](https://arxiv.org/pdf/2009.10053.pdf). We exclusively used the text from the [Corpus Corporum](https://www.mlat.uzh.ch) collected and maintained by the University of Zurich.
8
 
9
  The overall corpus contains 1.5G of text data (3x as much as has been used for V2 and very likely of better quality).
10