advaith5 commited on
Commit
48fac2a
·
verified ·
1 Parent(s): 811b39e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -18,7 +18,7 @@ tags:
18
 
19
  This model is an open-weights reranker model meant to be integrated into RAG applications to rerank results from preliminary search methods such as embeddings, BM25, and hybrid search.
20
 
21
- This reranker outperforms other popular rerankers such as coherer-rerank-v3.5 and Salesforce/Llama-rank-v1 across a wide variety of task domains, including on finance, legal, code, STEM, medical, and conversational data. See [this post](https://evals_blog_post) for more details.
22
  This model is trained on an innovative multi-stage pipeline that models query-document relevance scores using adjusted Elo-like ratings. See [this post](https://technical_blog_post) and our Technical Report (Coming soon!) for more details.
23
 
24
 
@@ -45,14 +45,14 @@ print(scores)
45
 
46
  Comparing NDCG@10 starting from top 100 documents by embedding (using text-3-embedding-small):
47
 
48
- | Task | Embedding | cohere-rerank-v3.5 | Salesforce/Llama-rank-v1 | zerank-1-small | zerank-1 |
49
- |----------------|-----------|--------------------|--------------------------|----------------|----------|
50
- | Code | 0.678 | 0.724 | 0.694 | 0.730 | 0.754 |
51
- | Conversational | 0.250 | 0.571 | 0.484 | 0.556 | 0.596 |
52
- | Finance | 0.839 | 0.824 | 0.828 | 0.861 | 0.894 |
53
- | Legal | 0.703 | 0.804 | 0.767 | 0.817 | 0.821 |
54
- | Medical | 0.619 | 0.750 | 0.719 | 0.773 | 0.796 |
55
- | STEM | 0.401 | 0.510 | 0.595 | 0.680 | 0.694 |
56
 
57
  Comparing BM25 and Hybrid Search without and with zerank-1:
58
 
 
18
 
19
  This model is an open-weights reranker model meant to be integrated into RAG applications to rerank results from preliminary search methods such as embeddings, BM25, and hybrid search.
20
 
21
+ This reranker outperforms other popular rerankers such as cohere-rerank-v3.5 and Salesforce/Llama-rank-v1 across a wide variety of task domains, including on finance, legal, code, STEM, medical, and conversational data. See [this post](https://evals_blog_post) for more details.
22
  This model is trained on an innovative multi-stage pipeline that models query-document relevance scores using adjusted Elo-like ratings. See [this post](https://technical_blog_post) and our Technical Report (Coming soon!) for more details.
23
 
24
 
 
45
 
46
  Comparing NDCG@10 starting from top 100 documents by embedding (using text-3-embedding-small):
47
 
48
+ | Task | Embedding | cohere-rerank-v3.5 | Salesforce/Llama-rank-v1 | zerank-1-small | **zerank-1** |
49
+ |----------------|-----------|--------------------|--------------------------|----------------|--------------|
50
+ | Code | 0.678 | 0.724 | 0.694 | 0.730 | **0.754** |
51
+ | Conversational | 0.250 | 0.571 | 0.484 | 0.556 | **0.596** |
52
+ | Finance | 0.839 | 0.824 | 0.828 | 0.861 | **0.894** |
53
+ | Legal | 0.703 | 0.804 | 0.767 | 0.817 | **0.821** |
54
+ | Medical | 0.619 | 0.750 | 0.719 | 0.773 | **0.796** |
55
+ | STEM | 0.401 | 0.510 | 0.595 | 0.680 | **0.694** |
56
 
57
  Comparing BM25 and Hybrid Search without and with zerank-1:
58