Vít Novotný commited on
Commit
12b87ed
·
1 Parent(s): 204c0b9

Document huggingface/transformers#16936 in `README.md`

Browse files
Files changed (1) hide show
  1. README.md +8 -1
README.md CHANGED
@@ -6,7 +6,7 @@ datasets:
6
  - math-stackexchange
7
  ---
8
 
9
- # MathBERTa base model
10
 
11
  Pretrained model on English language and LaTeX using a masked language modeling
12
  (MLM) objective. It was developed for [the ARQMath-3 shared task evaluation][1]
@@ -48,6 +48,11 @@ text generation you should look at model like GPT2.
48
 
49
  ### How to use
50
 
 
 
 
 
 
51
  You can use this model directly with a pipeline for masked language modeling:
52
 
53
  ```python
@@ -99,3 +104,5 @@ Together theses datasets weight 52GB of text and LaTeX.
99
 
100
  [5]: https://sigmathling.kwarc.info/resources/arxmliv-dataset-2020/
101
  [6]: https://www.cs.rit.edu/~dprl/ARQMath/arqmath-resources.html
 
 
 
6
  - math-stackexchange
7
  ---
8
 
9
+ # MathBERTa model
10
 
11
  Pretrained model on English language and LaTeX using a masked language modeling
12
  (MLM) objective. It was developed for [the ARQMath-3 shared task evaluation][1]
 
48
 
49
  ### How to use
50
 
51
+
52
+ *Due to the large number of added LaTeX tokens, MathBERTa is affected by [a
53
+ software bug in the 🤗 Transformers library][9] that causes it to load for tens
54
+ of minutes. The bug is [to be fixed in 🤗 Transformers 4.20.0][10].*
55
+
56
  You can use this model directly with a pipeline for masked language modeling:
57
 
58
  ```python
 
104
 
105
  [5]: https://sigmathling.kwarc.info/resources/arxmliv-dataset-2020/
106
  [6]: https://www.cs.rit.edu/~dprl/ARQMath/arqmath-resources.html
107
+ [9]: https://github.com/huggingface/transformers/issues/16936
108
+ [10]: https://github.com/huggingface/transformers/pull/17119