Update model card with paper details and GitHub link
#3
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,9 +1,10 @@
|
|
1 |
---
|
2 |
-
license: apache-2.0
|
3 |
-
language:
|
4 |
-
- en
|
5 |
base_model:
|
6 |
- Qwen/Qwen3-4B
|
|
|
|
|
|
|
|
|
7 |
pipeline_tag: text-ranking
|
8 |
tags:
|
9 |
- finance
|
@@ -11,14 +12,23 @@ tags:
|
|
11 |
- code
|
12 |
- stem
|
13 |
- medical
|
14 |
-
library_name: sentence-transformers
|
15 |
---
|
16 |
|
17 |
<img src="https://i.imgur.com/oxvhvQu.png"/>
|
18 |
|
19 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
-
In search
|
22 |
|
23 |
This 1.7B reranker is the smaller version of our flagship model [zeroentropy/zerank-1](https://huggingface.co/zeroentropy/zerank-1). Though the model is over 2x smaller, it maintains nearly the same standard of performance, continuing to outperform other popular rerankers, and displaying massive accuracy gains over traditional vector search.
|
24 |
|
|
|
1 |
---
|
|
|
|
|
|
|
2 |
base_model:
|
3 |
- Qwen/Qwen3-4B
|
4 |
+
language:
|
5 |
+
- en
|
6 |
+
library_name: sentence-transformers
|
7 |
+
license: apache-2.0
|
8 |
pipeline_tag: text-ranking
|
9 |
tags:
|
10 |
- finance
|
|
|
12 |
- code
|
13 |
- stem
|
14 |
- medical
|
|
|
15 |
---
|
16 |
|
17 |
<img src="https://i.imgur.com/oxvhvQu.png"/>
|
18 |
|
19 |
+
# zeroentropy/zerank-1-small
|
20 |
+
|
21 |
+
This model, `zeroentropy/zerank-1-small`, is a state-of-the-art open-weight reranker. It was introduced in the paper [zELO: ELO-inspired Training Method for Rerankers and Embedding Models](https://huggingface.co/papers/2509.12541).
|
22 |
+
|
23 |
+
## Paper: zELO: ELO-inspired Training Method for Rerankers and Embedding Models
|
24 |
+
|
25 |
+
### Abstract
|
26 |
+
We introduce a novel training methodology named zELO, which optimizes retrieval performance via the analysis that ranking tasks are statically equivalent to a Thurstone model. Based on the zELO method, we use unsupervised data in order train a suite of state-of-the-art open-weight reranker models: zerank-1 and zerank-1-small. These models achieve the highest retrieval scores in multiple domains, including finance, legal, code, and STEM, outperforming closed-source proprietary rerankers on both NDCG@10 and Recall. These models also demonstrate great versatility, maintaining their 0-shot performance on out-of-domain and private customer datasets. The training data included 112,000 queries and 100 documents per query, and was trained end-to-end from unannotated queries and documents in less than 10,000 H100-hours.
|
27 |
+
|
28 |
+
## Code
|
29 |
+
The methodology and benchmarking framework associated with this model can be found in the [zbench GitHub repository](https://github.com/zeroentropy-ai/zbench).
|
30 |
|
31 |
+
In search engines, [rerankers are crucial](https://www.zeroentropy.dev/blog/what-is-a-reranker-and-do-i-need-one) for improving the accuracy of your retrieval system.
|
32 |
|
33 |
This 1.7B reranker is the smaller version of our flagship model [zeroentropy/zerank-1](https://huggingface.co/zeroentropy/zerank-1). Though the model is over 2x smaller, it maintains nearly the same standard of performance, continuing to outperform other popular rerankers, and displaying massive accuracy gains over traditional vector search.
|
34 |
|