Improve model card
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,7 +1,11 @@
|
|
1 |
---
|
2 |
license: mit
|
|
|
|
|
3 |
---
|
4 |
|
5 |
-
This is the
|
6 |
-
|
7 |
-
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
+
pipeline_tag: text-classification
|
4 |
+
library_name: fasttext
|
5 |
---
|
6 |
|
7 |
+
This fastText model is a pretraining data filter, targeting the ARC Easy task. It's designed to select high-quality pretraining data using perplexity correlations, as described in [Improving Pretraining Data Using Perplexity Correlations](https://arxiv.org/abs/2409.05816). The model classifies text as either "include" or "exclude" for use in pretraining a language model. It does *not* itself represent a pretrained language model.
|
8 |
+
|
9 |
+
The filter was created using a method that leverages correlations between LLM losses on various texts and downstream benchmark performance. By selecting texts with high correlation, this model aims to improve the efficiency of the data selection process for pretraining LLMs.
|
10 |
+
|
11 |
+
Code: https://github.com/TristanThrush/perplexity-correlations
|