Vít Novotný
commited on
Commit
·
12b87ed
1
Parent(s):
204c0b9
Document huggingface/transformers#16936 in `README.md`
Browse files
README.md
CHANGED
@@ -6,7 +6,7 @@ datasets:
|
|
6 |
- math-stackexchange
|
7 |
---
|
8 |
|
9 |
-
# MathBERTa
|
10 |
|
11 |
Pretrained model on English language and LaTeX using a masked language modeling
|
12 |
(MLM) objective. It was developed for [the ARQMath-3 shared task evaluation][1]
|
@@ -48,6 +48,11 @@ text generation you should look at model like GPT2.
|
|
48 |
|
49 |
### How to use
|
50 |
|
|
|
|
|
|
|
|
|
|
|
51 |
You can use this model directly with a pipeline for masked language modeling:
|
52 |
|
53 |
```python
|
@@ -99,3 +104,5 @@ Together theses datasets weight 52GB of text and LaTeX.
|
|
99 |
|
100 |
[5]: https://sigmathling.kwarc.info/resources/arxmliv-dataset-2020/
|
101 |
[6]: https://www.cs.rit.edu/~dprl/ARQMath/arqmath-resources.html
|
|
|
|
|
|
6 |
- math-stackexchange
|
7 |
---
|
8 |
|
9 |
+
# MathBERTa model
|
10 |
|
11 |
Pretrained model on English language and LaTeX using a masked language modeling
|
12 |
(MLM) objective. It was developed for [the ARQMath-3 shared task evaluation][1]
|
|
|
48 |
|
49 |
### How to use
|
50 |
|
51 |
+
|
52 |
+
*Due to the large number of added LaTeX tokens, MathBERTa is affected by [a
|
53 |
+
software bug in the 🤗 Transformers library][9] that causes it to load for tens
|
54 |
+
of minutes. The bug is [to be fixed in 🤗 Transformers 4.20.0][10].*
|
55 |
+
|
56 |
You can use this model directly with a pipeline for masked language modeling:
|
57 |
|
58 |
```python
|
|
|
104 |
|
105 |
[5]: https://sigmathling.kwarc.info/resources/arxmliv-dataset-2020/
|
106 |
[6]: https://www.cs.rit.edu/~dprl/ARQMath/arqmath-resources.html
|
107 |
+
[9]: https://github.com/huggingface/transformers/issues/16936
|
108 |
+
[10]: https://github.com/huggingface/transformers/pull/17119
|