Update README.md
Browse files
README.md
CHANGED
|
@@ -17,13 +17,13 @@ tags:
|
|
| 17 |
|
| 18 |
# Model Card for Model ID
|
| 19 |
|
| 20 |
-
ModernBERT multi-task fine-tuned on tasksource NLI tasks, including MNLI, ANLI, SICK, WANLI, doc-nli, LingNLI, FOLIO, FOL-NLI, LogicNLI, Label-NLI and all datasets in the below table).
|
| 21 |
This is the equivalent of an "instruct" version.
|
| 22 |
The model was trained for 200k steps on an Nvidia A30 GPU.
|
| 23 |
|
| 24 |
It is very good at reasoning tasks (better than llama 3.1 8B Instruct on ANLI and FOLIO), long context reasoning, sentiment analysis and zero-shot classification with new labels.
|
| 25 |
|
| 26 |
-
The following table shows model test accuracy.
|
| 27 |
Further gains can be obtained by fine-tuning on a single-task, e.g. SST, but it this checkpoint is great for zero-shot classification and natural language inference (contradiction/entailment/neutral classification).
|
| 28 |
|
| 29 |
| test_name | test_accuracy |
|
|
|
|
| 17 |
|
| 18 |
# Model Card for Model ID
|
| 19 |
|
| 20 |
+
This model is ModernBERT multi-task fine-tuned on tasksource NLI tasks, including MNLI, ANLI, SICK, WANLI, doc-nli, LingNLI, FOLIO, FOL-NLI, LogicNLI, Label-NLI and all datasets in the below table).
|
| 21 |
This is the equivalent of an "instruct" version.
|
| 22 |
The model was trained for 200k steps on an Nvidia A30 GPU.
|
| 23 |
|
| 24 |
It is very good at reasoning tasks (better than llama 3.1 8B Instruct on ANLI and FOLIO), long context reasoning, sentiment analysis and zero-shot classification with new labels.
|
| 25 |
|
| 26 |
+
The following table shows model test accuracy. These are the scores for the same single transformer with different classification heads on top.
|
| 27 |
Further gains can be obtained by fine-tuning on a single-task, e.g. SST, but it this checkpoint is great for zero-shot classification and natural language inference (contradiction/entailment/neutral classification).
|
| 28 |
|
| 29 |
| test_name | test_accuracy |
|