Update model card with paper details and GitHub link

#3
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +16 -6
README.md CHANGED
@@ -1,9 +1,10 @@
1
  ---
2
- license: apache-2.0
3
- language:
4
- - en
5
  base_model:
6
  - Qwen/Qwen3-4B
 
 
 
 
7
  pipeline_tag: text-ranking
8
  tags:
9
  - finance
@@ -11,14 +12,23 @@ tags:
11
  - code
12
  - stem
13
  - medical
14
- library_name: sentence-transformers
15
  ---
16
 
17
  <img src="https://i.imgur.com/oxvhvQu.png"/>
18
 
19
- # Releasing zeroentropy/zerank-1-small
 
 
 
 
 
 
 
 
 
 
20
 
21
- In search enginers, [rerankers are crucial](https://www.zeroentropy.dev/blog/what-is-a-reranker-and-do-i-need-one) for improving the accuracy of your retrieval system.
22
 
23
  This 1.7B reranker is the smaller version of our flagship model [zeroentropy/zerank-1](https://huggingface.co/zeroentropy/zerank-1). Though the model is over 2x smaller, it maintains nearly the same standard of performance, continuing to outperform other popular rerankers, and displaying massive accuracy gains over traditional vector search.
24
 
 
1
  ---
 
 
 
2
  base_model:
3
  - Qwen/Qwen3-4B
4
+ language:
5
+ - en
6
+ library_name: sentence-transformers
7
+ license: apache-2.0
8
  pipeline_tag: text-ranking
9
  tags:
10
  - finance
 
12
  - code
13
  - stem
14
  - medical
 
15
  ---
16
 
17
  <img src="https://i.imgur.com/oxvhvQu.png"/>
18
 
19
+ # zeroentropy/zerank-1-small
20
+
21
+ This model, `zeroentropy/zerank-1-small`, is a state-of-the-art open-weight reranker. It was introduced in the paper [zELO: ELO-inspired Training Method for Rerankers and Embedding Models](https://huggingface.co/papers/2509.12541).
22
+
23
+ ## Paper: zELO: ELO-inspired Training Method for Rerankers and Embedding Models
24
+
25
+ ### Abstract
26
+ We introduce a novel training methodology named zELO, which optimizes retrieval performance via the analysis that ranking tasks are statically equivalent to a Thurstone model. Based on the zELO method, we use unsupervised data in order train a suite of state-of-the-art open-weight reranker models: zerank-1 and zerank-1-small. These models achieve the highest retrieval scores in multiple domains, including finance, legal, code, and STEM, outperforming closed-source proprietary rerankers on both NDCG@10 and Recall. These models also demonstrate great versatility, maintaining their 0-shot performance on out-of-domain and private customer datasets. The training data included 112,000 queries and 100 documents per query, and was trained end-to-end from unannotated queries and documents in less than 10,000 H100-hours.
27
+
28
+ ## Code
29
+ The methodology and benchmarking framework associated with this model can be found in the [zbench GitHub repository](https://github.com/zeroentropy-ai/zbench).
30
 
31
+ In search engines, [rerankers are crucial](https://www.zeroentropy.dev/blog/what-is-a-reranker-and-do-i-need-one) for improving the accuracy of your retrieval system.
32
 
33
  This 1.7B reranker is the smaller version of our flagship model [zeroentropy/zerank-1](https://huggingface.co/zeroentropy/zerank-1). Though the model is over 2x smaller, it maintains nearly the same standard of performance, continuing to outperform other popular rerankers, and displaying massive accuracy gains over traditional vector search.
34