|
eval/beir-arguana_ndcg@10 = 0.36665 |
|
eval/beir-arguana_recall@100 = 0.96728 |
|
eval/beir-avg_ndcg@10 = 0.30767 |
|
eval/beir-avg_recall@10 = 0.33533500000000005 |
|
eval/beir-avg_recall@100 = 0.523215 |
|
eval/beir-avg_recall@20 = 0.383875 |
|
eval/beir-climate-fever_ndcg@10 = 0.12105 |
|
eval/beir-climate-fever_recall@10 = 0.1548 |
|
eval/beir-climate-fever_recall@100 = 0.37817 |
|
eval/beir-climate-fever_recall@20 = 0.20165 |
|
eval/beir-cqadupstack_ndcg@10 = 0.3031158333333333 |
|
eval/beir-cqadupstack_recall@100 = 0.6389233333333334 |
|
eval/beir-dbpedia-entity_ndcg@10 = 0.3116 |
|
eval/beir-dbpedia-entity_recall@100 = 0.4652 |
|
eval/beir-fever_ndcg@10 = 0.63923 |
|
eval/beir-fever_recall@100 = 0.93145 |
|
eval/beir-fiqa_ndcg@10 = 0.28062 |
|
eval/beir-fiqa_recall@100 = 0.61765 |
|
eval/beir-hotpotqa_ndcg@10 = 0.49429 |
|
eval/beir-hotpotqa_recall@10 = 0.51587 |
|
eval/beir-hotpotqa_recall@100 = 0.66826 |
|
eval/beir-hotpotqa_recall@20 = 0.5661 |
|
eval/beir-msmarco_ndcg@10 = 0.20651 |
|
eval/beir-msmarco_recall@100 = 0.67481 |
|
eval/beir-nfcorpus_ndcg@10 = 0.34134 |
|
eval/beir-nfcorpus_recall@100 = 0.31962 |
|
eval/beir-nq_ndcg@10 = 0.27574 |
|
eval/beir-nq_recall@100 = 0.7927 |
|
eval/beir-quora_ndcg@10 = 0.78549 |
|
eval/beir-quora_recall@100 = 0.9752 |
|
eval/beir-scidocs_ndcg@10 = 0.15962 |
|
eval/beir-scidocs_recall@100 = 0.39283 |
|
eval/beir-scifact_ndcg@10 = 0.618 |
|
eval/beir-scifact_recall@100 = 0.911 |
|
eval/beir-trec-covid_ndcg@10 = 0.58244 |
|
eval/beir-trec-covid_recall@100 = 0.10597 |
|
eval/beir-webis-touche2020_ndcg@10 = 0.18565 |
|
eval/beir-webis-touche2020_recall@100 = 0.43843 |
|
eval/qa-curatedtrec-test-acc@100 = 0.9265129682997119 |
|
eval/qa-curatedtrec-test-acc@20 = 0.8371757925072046 |
|
eval/qa-curatedtrec-test-acc@5 = 0.6354466858789626 |
|
eval/qa-entityqs-macro-acc@100 = 0.7567939708496935 |
|
eval/qa-entityqs-macro-acc@20 = 0.6403777676502388 |
|
eval/qa-entityqs-macro-acc@5 = 0.49428236812619747 |
|
eval/qa-nq-test-acc@100 = 0.8024930747922437 |
|
eval/qa-nq-test-acc@20 = 0.6501385041551246 |
|
eval/qa-nq-test-acc@5 = 0.44155124653739614 |
|
eval/qa-squad1-test-acc@100 = 0.7376537369914853 |
|
eval/qa-squad1-test-acc@20 = 0.5704824976348155 |
|
eval/qa-squad1-test-acc@5 = 0.3709555345316935 |
|
eval/qa-trivia-test-acc@100 = 0.8229470520639972 |
|
eval/qa-trivia-test-acc@20 = 0.7227083885795103 |
|
eval/qa-trivia-test-acc@5 = 0.5689030319101919 |
|
eval/qa-webq-test-acc@100 = 0.8061023622047244 |
|
eval/qa-webq-test-acc@20 = 0.6520669291338582 |
|
eval/qa-webq-test-acc@5 = 0.40994094488188976 |
|
eval/senteval-CR = 87.02 |
|
eval/senteval-MPQA = 88.67 |
|
eval/senteval-MR = 80.75 |
|
eval/senteval-MRPC = 70.17 |
|
eval/senteval-SICKRelatedness = 0.7122486034563766 |
|
eval/senteval-SST2 = 82.8 |
|
eval/senteval-STS12 = 0.6520942930121344 |
|
eval/senteval-STS13 = 0.7489808796663362 |
|
eval/senteval-STS14 = 0.6214541243526541 |
|
eval/senteval-STS15 = 0.7497940183329284 |
|
eval/senteval-STS16 = 0.7490972681198437 |
|
eval/senteval-STSBenchmark = 0.708944371429387 |
|
eval/senteval-SUBJ = 95.02 |
|
eval/senteval-TREC = 78.63 |
|
eval/senteval-avg_sts_7 = 0.7060876511956659 |
|
eval/senteval-avg_transfer = 83.2942857142857 |
|
train/global_step = 100000 |
|
|