Update README.md
Browse files
README.md
CHANGED
@@ -18,7 +18,7 @@ tags:
|
|
18 |
|
19 |
This model is an open-weights reranker model meant to be integrated into RAG applications to rerank results from preliminary search methods such as embeddings, BM25, and hybrid search.
|
20 |
|
21 |
-
This reranker outperforms other popular rerankers such as
|
22 |
This model is trained on an innovative multi-stage pipeline that models query-document relevance scores using adjusted Elo-like ratings. See [this post](https://technical_blog_post) and our Technical Report (Coming soon!) for more details.
|
23 |
|
24 |
|
@@ -45,14 +45,14 @@ print(scores)
|
|
45 |
|
46 |
Comparing NDCG@10 starting from top 100 documents by embedding (using text-3-embedding-small):
|
47 |
|
48 |
-
| Task | Embedding | cohere-rerank-v3.5 | Salesforce/Llama-rank-v1 | zerank-1-small | zerank-1 |
|
49 |
-
|
50 |
-
| Code | 0.678 | 0.724 | 0.694 | 0.730 | 0.754 |
|
51 |
-
| Conversational | 0.250 | 0.571 | 0.484 | 0.556 | 0.596 |
|
52 |
-
| Finance | 0.839 | 0.824 | 0.828 | 0.861 | 0.894 |
|
53 |
-
| Legal | 0.703 | 0.804 | 0.767 | 0.817 | 0.821 |
|
54 |
-
| Medical | 0.619 | 0.750 | 0.719 | 0.773 | 0.796 |
|
55 |
-
| STEM | 0.401 | 0.510 | 0.595 | 0.680 | 0.694 |
|
56 |
|
57 |
Comparing BM25 and Hybrid Search without and with zerank-1:
|
58 |
|
|
|
18 |
|
19 |
This model is an open-weights reranker model meant to be integrated into RAG applications to rerank results from preliminary search methods such as embeddings, BM25, and hybrid search.
|
20 |
|
21 |
+
This reranker outperforms other popular rerankers such as cohere-rerank-v3.5 and Salesforce/Llama-rank-v1 across a wide variety of task domains, including on finance, legal, code, STEM, medical, and conversational data. See [this post](https://evals_blog_post) for more details.
|
22 |
This model is trained on an innovative multi-stage pipeline that models query-document relevance scores using adjusted Elo-like ratings. See [this post](https://technical_blog_post) and our Technical Report (Coming soon!) for more details.
|
23 |
|
24 |
|
|
|
45 |
|
46 |
Comparing NDCG@10 starting from top 100 documents by embedding (using text-3-embedding-small):
|
47 |
|
48 |
+
| Task | Embedding | cohere-rerank-v3.5 | Salesforce/Llama-rank-v1 | zerank-1-small | **zerank-1** |
|
49 |
+
|----------------|-----------|--------------------|--------------------------|----------------|--------------|
|
50 |
+
| Code | 0.678 | 0.724 | 0.694 | 0.730 | **0.754** |
|
51 |
+
| Conversational | 0.250 | 0.571 | 0.484 | 0.556 | **0.596** |
|
52 |
+
| Finance | 0.839 | 0.824 | 0.828 | 0.861 | **0.894** |
|
53 |
+
| Legal | 0.703 | 0.804 | 0.767 | 0.817 | **0.821** |
|
54 |
+
| Medical | 0.619 | 0.750 | 0.719 | 0.773 | **0.796** |
|
55 |
+
| STEM | 0.401 | 0.510 | 0.595 | 0.680 | **0.694** |
|
56 |
|
57 |
Comparing BM25 and Hybrid Search without and with zerank-1:
|
58 |
|