Roman Solomatin
commited on
add e5 instruct
Browse files- EXTERNAL_MODEL_RESULTS.json +697 -0
- model_meta.yaml +9 -0
EXTERNAL_MODEL_RESULTS.json
CHANGED
|
@@ -14202,6 +14202,703 @@
|
|
| 14202 |
]
|
| 14203 |
}
|
| 14204 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14205 |
"multilingual-e5-small": {
|
| 14206 |
"BitextMining": {
|
| 14207 |
"f1": [
|
|
|
|
| 14202 |
]
|
| 14203 |
}
|
| 14204 |
},
|
| 14205 |
+
"multilingual-e5-large-instruct": {
|
| 14206 |
+
"BitextMining": {
|
| 14207 |
+
"f1": [
|
| 14208 |
+
{
|
| 14209 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14210 |
+
"BornholmBitextMining": 55.42,
|
| 14211 |
+
"Tatoeba (rus-eng)": 93.7,
|
| 14212 |
+
"Tatoeba (spa-eng)": 98.53,
|
| 14213 |
+
"Tatoeba (isl-eng)": 95.2,
|
| 14214 |
+
"Tatoeba (jpn-eng)": 96.52,
|
| 14215 |
+
"Tatoeba (bre-eng)": 32.56,
|
| 14216 |
+
"Tatoeba (ast-eng)": 89.76,
|
| 14217 |
+
"Tatoeba (pms-eng)": 78.43,
|
| 14218 |
+
"Tatoeba (kab-eng)": 57.13,
|
| 14219 |
+
"Tatoeba (kat-eng)": 92.64,
|
| 14220 |
+
"Tatoeba (tur-eng)": 98.13,
|
| 14221 |
+
"Tatoeba (swh-eng)": 80.12,
|
| 14222 |
+
"Tatoeba (ukr-eng)": 94.8,
|
| 14223 |
+
"Tatoeba (epo-eng)": 98.38,
|
| 14224 |
+
"Tatoeba (lfn-eng)": 80.03,
|
| 14225 |
+
"Tatoeba (sqi-eng)": 96.45,
|
| 14226 |
+
"Tatoeba (cor-eng)": 11.54,
|
| 14227 |
+
"Tatoeba (pam-eng)": 19.75,
|
| 14228 |
+
"Tatoeba (gla-eng)": 77.87,
|
| 14229 |
+
"Tatoeba (amh-eng)": 87.64,
|
| 14230 |
+
"Tatoeba (jav-eng)": 85.29,
|
| 14231 |
+
"Tatoeba (lvs-eng)": 93.02,
|
| 14232 |
+
"Tatoeba (por-eng)": 94.91,
|
| 14233 |
+
"Tatoeba (arz-eng)": 82.26,
|
| 14234 |
+
"Tatoeba (aze-eng)": 93.89,
|
| 14235 |
+
"Tatoeba (yid-eng)": 90.47,
|
| 14236 |
+
"Tatoeba (gle-eng)": 84.37,
|
| 14237 |
+
"Tatoeba (pol-eng)": 97.33,
|
| 14238 |
+
"Tatoeba (ina-eng)": 96.1,
|
| 14239 |
+
"Tatoeba (ell-eng)": 94.85,
|
| 14240 |
+
"Tatoeba (fry-eng)": 80.83,
|
| 14241 |
+
"Tatoeba (kor-eng)": 91.77,
|
| 14242 |
+
"Tatoeba (ita-eng)": 94.37,
|
| 14243 |
+
"Tatoeba (ces-eng)": 96.23,
|
| 14244 |
+
"Tatoeba (fra-eng)": 94.97,
|
| 14245 |
+
"Tatoeba (ceb-eng)": 71.09,
|
| 14246 |
+
"Tatoeba (mal-eng)": 98.93,
|
| 14247 |
+
"Tatoeba (ido-eng)": 91.78,
|
| 14248 |
+
"Tatoeba (dtp-eng)": 14.1,
|
| 14249 |
+
"Tatoeba (kaz-eng)": 87.42,
|
| 14250 |
+
"Tatoeba (zsm-eng)": 96.62,
|
| 14251 |
+
"Tatoeba (dan-eng)": 95.87,
|
| 14252 |
+
"Tatoeba (nov-eng)": 83.42,
|
| 14253 |
+
"Tatoeba (tel-eng)": 95.51,
|
| 14254 |
+
"Tatoeba (arq-eng)": 64.1,
|
| 14255 |
+
"Tatoeba (dsb-eng)": 72.99,
|
| 14256 |
+
"Tatoeba (xho-eng)": 86.74,
|
| 14257 |
+
"Tatoeba (glg-eng)": 95.96,
|
| 14258 |
+
"Tatoeba (uig-eng)": 87.01,
|
| 14259 |
+
"Tatoeba (ang-eng)": 72.74,
|
| 14260 |
+
"Tatoeba (tzl-eng)": 59.17,
|
| 14261 |
+
"Tatoeba (deu-eng)": 99.33,
|
| 14262 |
+
"Tatoeba (ile-eng)": 89.94,
|
| 14263 |
+
"Tatoeba (lat-eng)": 74.17,
|
| 14264 |
+
"Tatoeba (eus-eng)": 84.67,
|
| 14265 |
+
"Tatoeba (bul-eng)": 94.52,
|
| 14266 |
+
"Tatoeba (hun-eng)": 94.52,
|
| 14267 |
+
"Tatoeba (mkd-eng)": 92.17,
|
| 14268 |
+
"Tatoeba (nds-eng)": 86.3,
|
| 14269 |
+
"Tatoeba (swg-eng)": 77.68,
|
| 14270 |
+
"Tatoeba (fao-eng)": 87.43,
|
| 14271 |
+
"Tatoeba (war-eng)": 73.83,
|
| 14272 |
+
"Tatoeba (csb-eng)": 68.19,
|
| 14273 |
+
"Tatoeba (afr-eng)": 95.67,
|
| 14274 |
+
"Tatoeba (bos-eng)": 95.86,
|
| 14275 |
+
"Tatoeba (cat-eng)": 95.53,
|
| 14276 |
+
"Tatoeba (hrv-eng)": 96.75,
|
| 14277 |
+
"Tatoeba (mhr-eng)": 16.98,
|
| 14278 |
+
"Tatoeba (tha-eng)": 96.59,
|
| 14279 |
+
"Tatoeba (tat-eng)": 85.7,
|
| 14280 |
+
"Tatoeba (oci-eng)": 72.35,
|
| 14281 |
+
"Tatoeba (urd-eng)": 93.22,
|
| 14282 |
+
"Tatoeba (kur-eng)": 77.3,
|
| 14283 |
+
"Tatoeba (lit-eng)": 91.71,
|
| 14284 |
+
"Tatoeba (cha-eng)": 50.55,
|
| 14285 |
+
"Tatoeba (ron-eng)": 97.32,
|
| 14286 |
+
"Tatoeba (pes-eng)": 94.92,
|
| 14287 |
+
"Tatoeba (cmn-eng)": 96.5,
|
| 14288 |
+
"Tatoeba (slv-eng)": 91.6,
|
| 14289 |
+
"Tatoeba (heb-eng)": 91.52,
|
| 14290 |
+
"Tatoeba (bel-eng)": 95.48,
|
| 14291 |
+
"Tatoeba (slk-eng)": 95.27,
|
| 14292 |
+
"Tatoeba (cbk-eng)": 83.24,
|
| 14293 |
+
"Tatoeba (yue-eng)": 92.84,
|
| 14294 |
+
"Tatoeba (hin-eng)": 97.47,
|
| 14295 |
+
"Tatoeba (nob-eng)": 98.33,
|
| 14296 |
+
"Tatoeba (gsw-eng)": 59.47,
|
| 14297 |
+
"Tatoeba (tgl-eng)": 96.0,
|
| 14298 |
+
"Tatoeba (nno-eng)": 94.06,
|
| 14299 |
+
"Tatoeba (est-eng)": 87.46,
|
| 14300 |
+
"Tatoeba (vie-eng)": 97.5,
|
| 14301 |
+
"Tatoeba (ber-eng)": 54.69,
|
| 14302 |
+
"Tatoeba (wuu-eng)": 92.4,
|
| 14303 |
+
"Tatoeba (tuk-eng)": 50.55,
|
| 14304 |
+
"Tatoeba (hye-eng)": 93.65,
|
| 14305 |
+
"Tatoeba (ara-eng)": 91.38,
|
| 14306 |
+
"Tatoeba (mon-eng)": 95.8,
|
| 14307 |
+
"Tatoeba (awa-eng)": 91.41,
|
| 14308 |
+
"Tatoeba (khm-eng)": 75.24,
|
| 14309 |
+
"Tatoeba (swe-eng)": 95.35,
|
| 14310 |
+
"Tatoeba (ind-eng)": 94.32,
|
| 14311 |
+
"Tatoeba (max-eng)": 74.56,
|
| 14312 |
+
"Tatoeba (hsb-eng)": 79.87,
|
| 14313 |
+
"Tatoeba (ben-eng)": 89.03,
|
| 14314 |
+
"Tatoeba (kzj-eng)": 14.96,
|
| 14315 |
+
"Tatoeba (uzb-eng)": 80.49,
|
| 14316 |
+
"Tatoeba (tam-eng)": 90.88,
|
| 14317 |
+
"Tatoeba (mar-eng)": 92.33,
|
| 14318 |
+
"Tatoeba (cym-eng)": 89.86,
|
| 14319 |
+
"Tatoeba (srp-eng)": 94.55,
|
| 14320 |
+
"Tatoeba (fin-eng)": 96.8,
|
| 14321 |
+
"Tatoeba (nld-eng)": 97.53,
|
| 14322 |
+
"Tatoeba (orv-eng)": 58.83
|
| 14323 |
+
}
|
| 14324 |
+
]
|
| 14325 |
+
},
|
| 14326 |
+
"Classification": {
|
| 14327 |
+
"accuracy": [
|
| 14328 |
+
{
|
| 14329 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14330 |
+
"AllegroReviews": 52.43,
|
| 14331 |
+
"AmazonCounterfactualClassification (en-ext)": 66.42,
|
| 14332 |
+
"AmazonCounterfactualClassification (en)": 69.72,
|
| 14333 |
+
"AmazonCounterfactualClassification (de)": 65.61,
|
| 14334 |
+
"AmazonCounterfactualClassification (ja)": 70.47,
|
| 14335 |
+
"AmazonPolarityClassification": 96.24,
|
| 14336 |
+
"AmazonReviewsClassification (en)": 56.2,
|
| 14337 |
+
"AmazonReviewsClassification (de)": 54.54,
|
| 14338 |
+
"AmazonReviewsClassification (es)": 49.88,
|
| 14339 |
+
"AmazonReviewsClassification (fr)": 49.78,
|
| 14340 |
+
"AmazonReviewsClassification (ja)": 49.36,
|
| 14341 |
+
"AmazonReviewsClassification (zh)": 45.1,
|
| 14342 |
+
"AngryTweetsClassification": 59.53,
|
| 14343 |
+
"Banking77Classification": 78.04,
|
| 14344 |
+
"CBD": 74.21,
|
| 14345 |
+
"DanishPoliticalCommentsClassification": 33.07,
|
| 14346 |
+
"EmotionClassification": 54.99,
|
| 14347 |
+
"GeoreviewClassification": 55.9,
|
| 14348 |
+
"HeadlineClassification": 86.18,
|
| 14349 |
+
"IFlyTek": 44.06,
|
| 14350 |
+
"ImdbClassification": 94.61,
|
| 14351 |
+
"InappropriatenessClassification": 65.53,
|
| 14352 |
+
"JDReview": 80.23,
|
| 14353 |
+
"KinopoiskClassification": 66.12,
|
| 14354 |
+
"LccSentimentClassification": 60.27,
|
| 14355 |
+
"MTOPDomainClassification (en)": 91.18,
|
| 14356 |
+
"MTOPDomainClassification (de)": 90.03,
|
| 14357 |
+
"MTOPDomainClassification (es)": 89.1,
|
| 14358 |
+
"MTOPDomainClassification (fr)": 85.89,
|
| 14359 |
+
"MTOPDomainClassification (hi)": 86.29,
|
| 14360 |
+
"MTOPDomainClassification (th)": 83.35,
|
| 14361 |
+
"MTOPIntentClassification (en)": 68.05,
|
| 14362 |
+
"MTOPIntentClassification (de)": 68.55,
|
| 14363 |
+
"MTOPIntentClassification (es)": 69.86,
|
| 14364 |
+
"MTOPIntentClassification (fr)": 63.26,
|
| 14365 |
+
"MTOPIntentClassification (hi)": 63.38,
|
| 14366 |
+
"MTOPIntentClassification (th)": 64.8,
|
| 14367 |
+
"MasakhaNEWSClassification (amh)": 88.48,
|
| 14368 |
+
"MasakhaNEWSClassification (eng)": 82.12,
|
| 14369 |
+
"MasakhaNEWSClassification (fra)": 78.93,
|
| 14370 |
+
"MasakhaNEWSClassification (hau)": 80.94,
|
| 14371 |
+
"MasakhaNEWSClassification (ibo)": 77.28,
|
| 14372 |
+
"MasakhaNEWSClassification (lin)": 75.94,
|
| 14373 |
+
"MasakhaNEWSClassification (lug)": 70.36,
|
| 14374 |
+
"MasakhaNEWSClassification (orm)": 80.28,
|
| 14375 |
+
"MasakhaNEWSClassification (pcm)": 93.74,
|
| 14376 |
+
"MasakhaNEWSClassification (run)": 83.88,
|
| 14377 |
+
"MasakhaNEWSClassification (sna)": 87.91,
|
| 14378 |
+
"MasakhaNEWSClassification (som)": 68.61,
|
| 14379 |
+
"MasakhaNEWSClassification (swa)": 79.47,
|
| 14380 |
+
"MasakhaNEWSClassification (tir)": 72.87,
|
| 14381 |
+
"MasakhaNEWSClassification (xho)": 84.95,
|
| 14382 |
+
"MasakhaNEWSClassification (yor)": 82.99,
|
| 14383 |
+
"MassiveIntentClassification (sl)": 63.79,
|
| 14384 |
+
"MassiveIntentClassification (cy)": 56.22,
|
| 14385 |
+
"MassiveIntentClassification (sq)": 62.12,
|
| 14386 |
+
"MassiveIntentClassification (ar)": 55.87,
|
| 14387 |
+
"MassiveIntentClassification (vi)": 64.49,
|
| 14388 |
+
"MassiveIntentClassification (fr)": 66.88,
|
| 14389 |
+
"MassiveIntentClassification (sw)": 54.14,
|
| 14390 |
+
"MassiveIntentClassification (es)": 65.56,
|
| 14391 |
+
"MassiveIntentClassification (az)": 61.96,
|
| 14392 |
+
"MassiveIntentClassification (hy)": 61.73,
|
| 14393 |
+
"MassiveIntentClassification (ru)": 67.6,
|
| 14394 |
+
"MassiveIntentClassification (tr)": 66.3,
|
| 14395 |
+
"MassiveIntentClassification (ta)": 60.4,
|
| 14396 |
+
"MassiveIntentClassification (lv)": 61.49,
|
| 14397 |
+
"MassiveIntentClassification (jv)": 55.6,
|
| 14398 |
+
"MassiveIntentClassification (fi)": 66.02,
|
| 14399 |
+
"MassiveIntentClassification (fa)": 68.48,
|
| 14400 |
+
"MassiveIntentClassification (ka)": 50.74,
|
| 14401 |
+
"MassiveIntentClassification (ro)": 63.73,
|
| 14402 |
+
"MassiveIntentClassification (sv)": 68.76,
|
| 14403 |
+
"MassiveIntentClassification (ms)": 63.88,
|
| 14404 |
+
"MassiveIntentClassification (kn)": 58.98,
|
| 14405 |
+
"MassiveIntentClassification (ml)": 63.22,
|
| 14406 |
+
"MassiveIntentClassification (pt)": 67.28,
|
| 14407 |
+
"MassiveIntentClassification (my)": 57.06,
|
| 14408 |
+
"MassiveIntentClassification (tl)": 60.99,
|
| 14409 |
+
"MassiveIntentClassification (af)": 60.95,
|
| 14410 |
+
"MassiveIntentClassification (he)": 63.5,
|
| 14411 |
+
"MassiveIntentClassification (ja)": 68.29,
|
| 14412 |
+
"MassiveIntentClassification (zh-CN)": 67.6,
|
| 14413 |
+
"MassiveIntentClassification (km)": 46.88,
|
| 14414 |
+
"MassiveIntentClassification (mn)": 58.85,
|
| 14415 |
+
"MassiveIntentClassification (am)": 54.28,
|
| 14416 |
+
"MassiveIntentClassification (de)": 65.91,
|
| 14417 |
+
"MassiveIntentClassification (el)": 65.25,
|
| 14418 |
+
"MassiveIntentClassification (hu)": 65.54,
|
| 14419 |
+
"MassiveIntentClassification (en)": 70.89,
|
| 14420 |
+
"MassiveIntentClassification (id)": 65.45,
|
| 14421 |
+
"MassiveIntentClassification (ur)": 60.74,
|
| 14422 |
+
"MassiveIntentClassification (nl)": 68.18,
|
| 14423 |
+
"MassiveIntentClassification (da)": 65.08,
|
| 14424 |
+
"MassiveIntentClassification (th)": 62.58,
|
| 14425 |
+
"MassiveIntentClassification (pl)": 67.45,
|
| 14426 |
+
"MassiveIntentClassification (zh-TW)": 62.09,
|
| 14427 |
+
"MassiveIntentClassification (bn)": 61.73,
|
| 14428 |
+
"MassiveIntentClassification (is)": 57.29,
|
| 14429 |
+
"MassiveIntentClassification (te)": 61.38,
|
| 14430 |
+
"MassiveIntentClassification (ko)": 64.16,
|
| 14431 |
+
"MassiveIntentClassification (nb)": 65.85,
|
| 14432 |
+
"MassiveIntentClassification (it)": 66.82,
|
| 14433 |
+
"MassiveIntentClassification (hi)": 65.27,
|
| 14434 |
+
"MassiveScenarioClassification (ar)": 63.86,
|
| 14435 |
+
"MassiveScenarioClassification (id)": 70.52,
|
| 14436 |
+
"MassiveScenarioClassification (sl)": 68.6,
|
| 14437 |
+
"MassiveScenarioClassification (sq)": 69.09,
|
| 14438 |
+
"MassiveScenarioClassification (km)": 54.17,
|
| 14439 |
+
"MassiveScenarioClassification (fr)": 71.16,
|
| 14440 |
+
"MassiveScenarioClassification (bn)": 67.55,
|
| 14441 |
+
"MassiveScenarioClassification (am)": 61.43,
|
| 14442 |
+
"MassiveScenarioClassification (nl)": 73.21,
|
| 14443 |
+
"MassiveScenarioClassification (af)": 68.3,
|
| 14444 |
+
"MassiveScenarioClassification (fi)": 68.99,
|
| 14445 |
+
"MassiveScenarioClassification (ja)": 73.43,
|
| 14446 |
+
"MassiveScenarioClassification (en)": 73.88,
|
| 14447 |
+
"MassiveScenarioClassification (pt)": 69.76,
|
| 14448 |
+
"MassiveScenarioClassification (ka)": 59.88,
|
| 14449 |
+
"MassiveScenarioClassification (da)": 71.88,
|
| 14450 |
+
"MassiveScenarioClassification (jv)": 63.3,
|
| 14451 |
+
"MassiveScenarioClassification (zh-TW)": 68.04,
|
| 14452 |
+
"MassiveScenarioClassification (cy)": 62.98,
|
| 14453 |
+
"MassiveScenarioClassification (hu)": 71.72,
|
| 14454 |
+
"MassiveScenarioClassification (lv)": 67.54,
|
| 14455 |
+
"MassiveScenarioClassification (pl)": 71.44,
|
| 14456 |
+
"MassiveScenarioClassification (de)": 72.68,
|
| 14457 |
+
"MassiveScenarioClassification (ms)": 69.07,
|
| 14458 |
+
"MassiveScenarioClassification (ur)": 67.11,
|
| 14459 |
+
"MassiveScenarioClassification (kn)": 66.84,
|
| 14460 |
+
"MassiveScenarioClassification (ta)": 64.96,
|
| 14461 |
+
"MassiveScenarioClassification (is)": 67.14,
|
| 14462 |
+
"MassiveScenarioClassification (it)": 70.44,
|
| 14463 |
+
"MassiveScenarioClassification (sv)": 73.38,
|
| 14464 |
+
"MassiveScenarioClassification (te)": 67.53,
|
| 14465 |
+
"MassiveScenarioClassification (th)": 69.54,
|
| 14466 |
+
"MassiveScenarioClassification (fa)": 72.17,
|
| 14467 |
+
"MassiveScenarioClassification (mn)": 63.21,
|
| 14468 |
+
"MassiveScenarioClassification (ro)": 68.72,
|
| 14469 |
+
"MassiveScenarioClassification (ml)": 69.13,
|
| 14470 |
+
"MassiveScenarioClassification (az)": 66.18,
|
| 14471 |
+
"MassiveScenarioClassification (hi)": 69.83,
|
| 14472 |
+
"MassiveScenarioClassification (el)": 71.1,
|
| 14473 |
+
"MassiveScenarioClassification (nb)": 71.28,
|
| 14474 |
+
"MassiveScenarioClassification (tl)": 67.64,
|
| 14475 |
+
"MassiveScenarioClassification (he)": 67.63,
|
| 14476 |
+
"MassiveScenarioClassification (vi)": 69.17,
|
| 14477 |
+
"MassiveScenarioClassification (es)": 70.0,
|
| 14478 |
+
"MassiveScenarioClassification (ko)": 70.49,
|
| 14479 |
+
"MassiveScenarioClassification (tr)": 69.63,
|
| 14480 |
+
"MassiveScenarioClassification (ru)": 71.59,
|
| 14481 |
+
"MassiveScenarioClassification (zh-CN)": 72.41,
|
| 14482 |
+
"MassiveScenarioClassification (my)": 63.12,
|
| 14483 |
+
"MassiveScenarioClassification (sw)": 62.96,
|
| 14484 |
+
"MassiveScenarioClassification (hy)": 65.85,
|
| 14485 |
+
"MultilingualSentiment": 71.66,
|
| 14486 |
+
"NoRecClassification": 57.06,
|
| 14487 |
+
"NordicLangClassification": 76.57,
|
| 14488 |
+
"OnlineShopping": 91.83,
|
| 14489 |
+
"PAC": 65.68,
|
| 14490 |
+
"PolEmo2.0-IN": 80.96,
|
| 14491 |
+
"PolEmo2.0-OUT": 56.96,
|
| 14492 |
+
"RuReviewsClassification": 68.56,
|
| 14493 |
+
"RuSciBenchGRNTIClassification": 65.07,
|
| 14494 |
+
"RuSciBenchOECDClassification": 50.21,
|
| 14495 |
+
"TNews": 49.85,
|
| 14496 |
+
"ToxicConversationsClassification": 66.82,
|
| 14497 |
+
"TweetSentimentExtractionClassification": 59.18,
|
| 14498 |
+
"Waimai": 86.83
|
| 14499 |
+
}
|
| 14500 |
+
]
|
| 14501 |
+
},
|
| 14502 |
+
"Clustering": {
|
| 14503 |
+
"v_measure": [
|
| 14504 |
+
{
|
| 14505 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14506 |
+
"GeoreviewClusteringP2P": 74.34,
|
| 14507 |
+
"MLSUMClusteringP2P (ru)": 57.77,
|
| 14508 |
+
"MLSUMClusteringS2S (ru)": 57.5,
|
| 14509 |
+
"MasakhaNEWSClusteringP2P (amh)": 74.82,
|
| 14510 |
+
"MasakhaNEWSClusteringP2P (eng)": 70.12,
|
| 14511 |
+
"MasakhaNEWSClusteringP2P (fra)": 70.48,
|
| 14512 |
+
"MasakhaNEWSClusteringP2P (hau)": 71.22,
|
| 14513 |
+
"MasakhaNEWSClusteringP2P (ibo)": 56.64,
|
| 14514 |
+
"MasakhaNEWSClusteringP2P (lin)": 72.94,
|
| 14515 |
+
"MasakhaNEWSClusteringP2P (lug)": 65.94,
|
| 14516 |
+
"MasakhaNEWSClusteringP2P (orm)": 45.85,
|
| 14517 |
+
"MasakhaNEWSClusteringP2P (pcm)": 88.57,
|
| 14518 |
+
"MasakhaNEWSClusteringP2P (run)": 64.64,
|
| 14519 |
+
"MasakhaNEWSClusteringP2P (sna)": 81.13,
|
| 14520 |
+
"MasakhaNEWSClusteringP2P (som)": 54.59,
|
| 14521 |
+
"MasakhaNEWSClusteringP2P (swa)": 40.55,
|
| 14522 |
+
"MasakhaNEWSClusteringP2P (tir)": 62.76,
|
| 14523 |
+
"MasakhaNEWSClusteringP2P (xho)": 57.61,
|
| 14524 |
+
"MasakhaNEWSClusteringP2P (yor)": 64.78,
|
| 14525 |
+
"MasakhaNEWSClusteringS2S (amh)": 67.38,
|
| 14526 |
+
"MasakhaNEWSClusteringS2S (eng)": 61.2,
|
| 14527 |
+
"MasakhaNEWSClusteringS2S (fra)": 69.48,
|
| 14528 |
+
"MasakhaNEWSClusteringS2S (hau)": 52.67,
|
| 14529 |
+
"MasakhaNEWSClusteringS2S (ibo)": 78.68,
|
| 14530 |
+
"MasakhaNEWSClusteringS2S (lin)": 80.26,
|
| 14531 |
+
"MasakhaNEWSClusteringS2S (lug)": 49.73,
|
| 14532 |
+
"MasakhaNEWSClusteringS2S (orm)": 44.39,
|
| 14533 |
+
"MasakhaNEWSClusteringS2S (pcm)": 74.0,
|
| 14534 |
+
"MasakhaNEWSClusteringS2S (run)": 64.34,
|
| 14535 |
+
"MasakhaNEWSClusteringS2S (sna)": 67.1,
|
| 14536 |
+
"MasakhaNEWSClusteringS2S (som)": 44.64,
|
| 14537 |
+
"MasakhaNEWSClusteringS2S (swa)": 42.09,
|
| 14538 |
+
"MasakhaNEWSClusteringS2S (tir)": 54.35,
|
| 14539 |
+
"MasakhaNEWSClusteringS2S (xho)": 43.19,
|
| 14540 |
+
"MasakhaNEWSClusteringS2S (yor)": 54.23,
|
| 14541 |
+
"RuSciBenchGRNTIClusteringP2P": 62.21,
|
| 14542 |
+
"RuSciBenchOECDClusteringP2P": 53.09,
|
| 14543 |
+
"TwentyNewsgroupsClustering": 51.03
|
| 14544 |
+
}
|
| 14545 |
+
]
|
| 14546 |
+
},
|
| 14547 |
+
"PairClassification": {
|
| 14548 |
+
"max_ap": [
|
| 14549 |
+
{
|
| 14550 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14551 |
+
"CDSC-E": 76.17,
|
| 14552 |
+
"OpusparcusPC (de)": 97.56,
|
| 14553 |
+
"OpusparcusPC (en)": 98.91,
|
| 14554 |
+
"OpusparcusPC (fi)": 94.74,
|
| 14555 |
+
"OpusparcusPC (fr)": 94.73,
|
| 14556 |
+
"OpusparcusPC (ru)": 91.1,
|
| 14557 |
+
"OpusparcusPC (sv)": 95.78,
|
| 14558 |
+
"PSC": 99.31,
|
| 14559 |
+
"PawsXPairClassification (de)": 56.46,
|
| 14560 |
+
"PawsXPairClassification (en)": 64.4,
|
| 14561 |
+
"PawsXPairClassification (es)": 57.27,
|
| 14562 |
+
"PawsXPairClassification (fr)": 59.73,
|
| 14563 |
+
"PawsXPairClassification (ja)": 50.82,
|
| 14564 |
+
"PawsXPairClassification (ko)": 51.63,
|
| 14565 |
+
"PawsXPairClassification (zh)": 57.34,
|
| 14566 |
+
"SICK-E-PL": 79.69,
|
| 14567 |
+
"SprintDuplicateQuestions": 92.19,
|
| 14568 |
+
"TERRa": 63.77,
|
| 14569 |
+
"TwitterSemEval2015": 79.79,
|
| 14570 |
+
"TwitterURLCorpus": 86.73
|
| 14571 |
+
},
|
| 14572 |
+
{
|
| 14573 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14574 |
+
"CDSC-E": 76.17,
|
| 14575 |
+
"OpusparcusPC (de)": 97.56,
|
| 14576 |
+
"OpusparcusPC (en)": 98.92,
|
| 14577 |
+
"OpusparcusPC (fi)": 94.74,
|
| 14578 |
+
"OpusparcusPC (fr)": 94.73,
|
| 14579 |
+
"OpusparcusPC (ru)": 91.1,
|
| 14580 |
+
"OpusparcusPC (sv)": 95.78,
|
| 14581 |
+
"PSC": 99.31,
|
| 14582 |
+
"PawsXPairClassification (de)": 56.64,
|
| 14583 |
+
"PawsXPairClassification (en)": 64.43,
|
| 14584 |
+
"PawsXPairClassification (es)": 57.28,
|
| 14585 |
+
"PawsXPairClassification (fr)": 59.76,
|
| 14586 |
+
"PawsXPairClassification (ja)": 50.82,
|
| 14587 |
+
"PawsXPairClassification (ko)": 51.74,
|
| 14588 |
+
"PawsXPairClassification (zh)": 57.75,
|
| 14589 |
+
"SICK-E-PL": 79.69,
|
| 14590 |
+
"SprintDuplicateQuestions": 92.19,
|
| 14591 |
+
"TERRa": 63.89,
|
| 14592 |
+
"TwitterSemEval2015": 79.79,
|
| 14593 |
+
"TwitterURLCorpus": 86.73
|
| 14594 |
+
}
|
| 14595 |
+
]
|
| 14596 |
+
},
|
| 14597 |
+
"Reranking": {
|
| 14598 |
+
"map": [
|
| 14599 |
+
{
|
| 14600 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14601 |
+
"AlloprofReranking": 74.68,
|
| 14602 |
+
"AskUbuntuDupQuestions": 64.41,
|
| 14603 |
+
"MMarcoReranking": 23.6,
|
| 14604 |
+
"MindSmallReranking": 33.07,
|
| 14605 |
+
"RuBQReranking": 75.84,
|
| 14606 |
+
"SciDocsRR": 85.75,
|
| 14607 |
+
"StackOverflowDupQuestions": 52.45,
|
| 14608 |
+
"SyntecReranking": 89.95,
|
| 14609 |
+
"T2Reranking": 67.12
|
| 14610 |
+
},
|
| 14611 |
+
{
|
| 14612 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14613 |
+
"MIRACLReranking (ru)": 62.49
|
| 14614 |
+
}
|
| 14615 |
+
]
|
| 14616 |
+
},
|
| 14617 |
+
"Retrieval": {
|
| 14618 |
+
"ndcg_at_10": [
|
| 14619 |
+
{
|
| 14620 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14621 |
+
"AILACasedocs": 33.33,
|
| 14622 |
+
"AILAStatutes": 29.66,
|
| 14623 |
+
"ARCChallenge": 15.03,
|
| 14624 |
+
"AlloprofRetrieval": 52.12,
|
| 14625 |
+
"AlphaNLI": 24.87,
|
| 14626 |
+
"AppsRetrieval": 34.89,
|
| 14627 |
+
"ArguAna": 58.48,
|
| 14628 |
+
"ArguAna-PL": 44.6,
|
| 14629 |
+
"BSARDRetrieval": 24.61,
|
| 14630 |
+
"CmedqaRetrieval": 34.15,
|
| 14631 |
+
"CodeFeedbackMT": 39.87,
|
| 14632 |
+
"CodeFeedbackST": 75.89,
|
| 14633 |
+
"CodeSearchNetCCRetrieval (python)": 86.24,
|
| 14634 |
+
"CodeSearchNetCCRetrieval (javascript)": 79.2,
|
| 14635 |
+
"CodeSearchNetCCRetrieval (go)": 72.31,
|
| 14636 |
+
"CodeSearchNetCCRetrieval (ruby)": 81.9,
|
| 14637 |
+
"CodeSearchNetCCRetrieval (java)": 80.07,
|
| 14638 |
+
"CodeSearchNetCCRetrieval (php)": 72.49,
|
| 14639 |
+
"CodeSearchNetRetrieval (python)": 88.96,
|
| 14640 |
+
"CodeSearchNetRetrieval (javascript)": 75.37,
|
| 14641 |
+
"CodeSearchNetRetrieval (go)": 90.41,
|
| 14642 |
+
"CodeSearchNetRetrieval (ruby)": 79.56,
|
| 14643 |
+
"CodeSearchNetRetrieval (java)": 80.55,
|
| 14644 |
+
"CodeSearchNetRetrieval (php)": 83.7,
|
| 14645 |
+
"CodeTransOceanContest": 83.36,
|
| 14646 |
+
"CodeTransOceanDL": 29.01,
|
| 14647 |
+
"CosQA": 37.75,
|
| 14648 |
+
"CovidRetrieval": 75.8,
|
| 14649 |
+
"DuRetrieval": 85.16,
|
| 14650 |
+
"EcomRetrieval": 53.92,
|
| 14651 |
+
"FiQA-PL": 32.01,
|
| 14652 |
+
"FiQA2018": 48.42,
|
| 14653 |
+
"GerDaLIRSmall": 21.34,
|
| 14654 |
+
"HellaSwag": 32.02,
|
| 14655 |
+
"LEMBNarrativeQARetrieval": 26.71,
|
| 14656 |
+
"LEMBQMSumRetrieval": 26.08,
|
| 14657 |
+
"LEMBSummScreenFDRetrieval": 72.75,
|
| 14658 |
+
"LEMBWikimQARetrieval": 57.79,
|
| 14659 |
+
"LeCaRDv2": 64.22,
|
| 14660 |
+
"LegalBenchConsumerContractsQA": 76.8,
|
| 14661 |
+
"LegalBenchCorporateLobbying": 94.25,
|
| 14662 |
+
"LegalQuAD": 51.25,
|
| 14663 |
+
"LegalSummarization": 68.07,
|
| 14664 |
+
"MIRACLRetrieval (ru)": 66.08,
|
| 14665 |
+
"MMarcoRetrieval": 78.81,
|
| 14666 |
+
"MedicalRetrieval": 56.55,
|
| 14667 |
+
"MintakaRetrieval (ar)": 26.13,
|
| 14668 |
+
"MintakaRetrieval (de)": 35.13,
|
| 14669 |
+
"MintakaRetrieval (es)": 34.51,
|
| 14670 |
+
"MintakaRetrieval (fr)": 33.49,
|
| 14671 |
+
"MintakaRetrieval (hi)": 28.19,
|
| 14672 |
+
"MintakaRetrieval (it)": 35.01,
|
| 14673 |
+
"MintakaRetrieval (ja)": 26.31,
|
| 14674 |
+
"MintakaRetrieval (pt)": 35.11,
|
| 14675 |
+
"NFCorpus": 36.34,
|
| 14676 |
+
"NFCorpus-PL": 32.66,
|
| 14677 |
+
"PIQA": 32.26,
|
| 14678 |
+
"Quail": 8.63,
|
| 14679 |
+
"RARbCode": 71.22,
|
| 14680 |
+
"RARbMath": 71.95,
|
| 14681 |
+
"RiaNewsRetrieval": 83.26,
|
| 14682 |
+
"RuBQRetrieval": 73.9,
|
| 14683 |
+
"SCIDOCS": 19.24,
|
| 14684 |
+
"SCIDOCS-PL": 17.15,
|
| 14685 |
+
"SIQA": 7.33,
|
| 14686 |
+
"SciFact": 71.62,
|
| 14687 |
+
"SciFact-PL": 63.31,
|
| 14688 |
+
"SpartQA": 13.54,
|
| 14689 |
+
"StackOverflowQA": 85.82,
|
| 14690 |
+
"SyntecRetrieval": 87.8,
|
| 14691 |
+
"SyntheticText2SQL": 58.85,
|
| 14692 |
+
"T2Retrieval": 82.92,
|
| 14693 |
+
"TRECCOVID": 82.51,
|
| 14694 |
+
"TRECCOVID-PL": 62.03,
|
| 14695 |
+
"TempReasonL1": 1.2,
|
| 14696 |
+
"TempReasonL2Fact": 40.19,
|
| 14697 |
+
"TempReasonL2Pure": 3.64,
|
| 14698 |
+
"TempReasonL3Fact": 37.37,
|
| 14699 |
+
"TempReasonL3Pure": 9.82,
|
| 14700 |
+
"Touche2020": 27.4,
|
| 14701 |
+
"VideoRetrieval": 52.24,
|
| 14702 |
+
"WinoGrande": 54.27,
|
| 14703 |
+
"XPQARetrieval (ara-ara)": 48.56,
|
| 14704 |
+
"XPQARetrieval (eng-ara)": 34.01,
|
| 14705 |
+
"XPQARetrieval (ara-eng)": 45.13,
|
| 14706 |
+
"XPQARetrieval (deu-deu)": 79.18,
|
| 14707 |
+
"XPQARetrieval (eng-deu)": 50.55,
|
| 14708 |
+
"XPQARetrieval (deu-eng)": 75.23,
|
| 14709 |
+
"XPQARetrieval (spa-spa)": 65.07,
|
| 14710 |
+
"XPQARetrieval (eng-spa)": 41.75,
|
| 14711 |
+
"XPQARetrieval (spa-eng)": 61.02,
|
| 14712 |
+
"XPQARetrieval (fra-fra)": 72.72,
|
| 14713 |
+
"XPQARetrieval (eng-fra)": 46.17,
|
| 14714 |
+
"XPQARetrieval (fra-eng)": 69.33,
|
| 14715 |
+
"XPQARetrieval (hin-hin)": 76.62,
|
| 14716 |
+
"XPQARetrieval (eng-hin)": 18.12,
|
| 14717 |
+
"XPQARetrieval (hin-eng)": 71.95,
|
| 14718 |
+
"XPQARetrieval (ita-ita)": 77.03,
|
| 14719 |
+
"XPQARetrieval (eng-ita)": 41.42,
|
| 14720 |
+
"XPQARetrieval (ita-eng)": 70.6,
|
| 14721 |
+
"XPQARetrieval (jpn-jpn)": 74.36,
|
| 14722 |
+
"XPQARetrieval (eng-jpn)": 41.97,
|
| 14723 |
+
"XPQARetrieval (jpn-eng)": 70.5,
|
| 14724 |
+
"XPQARetrieval (kor-kor)": 39.74,
|
| 14725 |
+
"XPQARetrieval (eng-kor)": 34.9,
|
| 14726 |
+
"XPQARetrieval (kor-eng)": 37.39,
|
| 14727 |
+
"XPQARetrieval (pol-pol)": 52.53,
|
| 14728 |
+
"XPQARetrieval (eng-pol)": 34.22,
|
| 14729 |
+
"XPQARetrieval (pol-eng)": 48.53,
|
| 14730 |
+
"XPQARetrieval (por-por)": 51.17,
|
| 14731 |
+
"XPQARetrieval (eng-por)": 31.76,
|
| 14732 |
+
"XPQARetrieval (por-eng)": 48.79,
|
| 14733 |
+
"XPQARetrieval (tam-tam)": 49.69,
|
| 14734 |
+
"XPQARetrieval (eng-tam)": 6.8,
|
| 14735 |
+
"XPQARetrieval (tam-eng)": 40.83,
|
| 14736 |
+
"XPQARetrieval (cmn-cmn)": 66.42,
|
| 14737 |
+
"XPQARetrieval (eng-cmn)": 33.94,
|
| 14738 |
+
"XPQARetrieval (cmn-eng)": 59.78
|
| 14739 |
+
}
|
| 14740 |
+
]
|
| 14741 |
+
},
|
| 14742 |
+
"STS": {
|
| 14743 |
+
"cosine_spearman": [
|
| 14744 |
+
{
|
| 14745 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14746 |
+
"AFQMC": 37.53,
|
| 14747 |
+
"ATEC": 43.27,
|
| 14748 |
+
"BIOSSES": 87.46,
|
| 14749 |
+
"BQ": 48.8,
|
| 14750 |
+
"CDSC-R": 92.35,
|
| 14751 |
+
"LCQMC": 76.06,
|
| 14752 |
+
"PAWSX": 15.06,
|
| 14753 |
+
"RUParaPhraserSTS": 75.4,
|
| 14754 |
+
"RuSTSBenchmarkSTS": 83.97,
|
| 14755 |
+
"SICK-R": 81.73,
|
| 14756 |
+
"SICK-R-PL": 77.62,
|
| 14757 |
+
"SICKFr": 80.2,
|
| 14758 |
+
"STS12": 82.53,
|
| 14759 |
+
"STS13": 88.05,
|
| 14760 |
+
"STS14": 84.83,
|
| 14761 |
+
"STS15": 91.02,
|
| 14762 |
+
"STS16": 87.32,
|
| 14763 |
+
"STS17 (ar-ar)": 82.71,
|
| 14764 |
+
"STS17 (nl-en)": 87.44,
|
| 14765 |
+
"STS17 (fr-en)": 86.28,
|
| 14766 |
+
"STS17 (ko-ko)": 84.31,
|
| 14767 |
+
"STS17 (es-es)": 89.4,
|
| 14768 |
+
"STS17 (en-ar)": 81.83,
|
| 14769 |
+
"STS17 (en-de)": 87.61,
|
| 14770 |
+
"STS17 (en-tr)": 80.47,
|
| 14771 |
+
"STS17 (es-en)": 87.03,
|
| 14772 |
+
"STS17 (en-en)": 90.33,
|
| 14773 |
+
"STS17 (it-en)": 88.25,
|
| 14774 |
+
"STS22 (es-en)": 78.65,
|
| 14775 |
+
"STS22 (de)": 61.53,
|
| 14776 |
+
"STS22 (zh)": 67.4,
|
| 14777 |
+
"STS22 (es)": 68.45,
|
| 14778 |
+
"STS22 (de-fr)": 65.52,
|
| 14779 |
+
"STS22 (pl)": 40.97,
|
| 14780 |
+
"STS22 (ru)": 65.17,
|
| 14781 |
+
"STS22 (en)": 68.67,
|
| 14782 |
+
"STS22 (fr)": 82.25,
|
| 14783 |
+
"STS22 (es-it)": 75.25,
|
| 14784 |
+
"STS22 (pl-en)": 74.86,
|
| 14785 |
+
"STS22 (fr-pl)": 61.98,
|
| 14786 |
+
"STS22 (ar)": 59.52,
|
| 14787 |
+
"STS22 (it)": 79.45,
|
| 14788 |
+
"STS22 (tr)": 67.2,
|
| 14789 |
+
"STS22 (zh-en)": 71.45,
|
| 14790 |
+
"STS22 (de-pl)": 54.6,
|
| 14791 |
+
"STS22 (de-en)": 58.22,
|
| 14792 |
+
"STSB": 81.68,
|
| 14793 |
+
"STSBenchmark": 88.39,
|
| 14794 |
+
"STSBenchmarkMultilingualSTS (de)": 85.38,
|
| 14795 |
+
"STSBenchmarkMultilingualSTS (nl)": 84.29,
|
| 14796 |
+
"STSBenchmarkMultilingualSTS (pl)": 83.42,
|
| 14797 |
+
"STSBenchmarkMultilingualSTS (fr)": 84.94,
|
| 14798 |
+
"STSBenchmarkMultilingualSTS (pt)": 84.9,
|
| 14799 |
+
"STSBenchmarkMultilingualSTS (it)": 84.38,
|
| 14800 |
+
"STSBenchmarkMultilingualSTS (es)": 86.15,
|
| 14801 |
+
"STSBenchmarkMultilingualSTS (zh)": 82.54,
|
| 14802 |
+
"STSBenchmarkMultilingualSTS (en)": 88.39,
|
| 14803 |
+
"STSBenchmarkMultilingualSTS (ru)": 83.86
|
| 14804 |
+
},
|
| 14805 |
+
{
|
| 14806 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14807 |
+
"AFQMC": 37.53,
|
| 14808 |
+
"ATEC": 43.27,
|
| 14809 |
+
"BIOSSES": 87.46,
|
| 14810 |
+
"BQ": 48.8,
|
| 14811 |
+
"CDSC-R": 92.35,
|
| 14812 |
+
"LCQMC": 76.06,
|
| 14813 |
+
"PAWSX": 15.06,
|
| 14814 |
+
"RUParaPhraserSTS": 75.4,
|
| 14815 |
+
"RuSTSBenchmarkSTS": 83.97,
|
| 14816 |
+
"SICK-R": 81.73,
|
| 14817 |
+
"SICK-R-PL": 77.62,
|
| 14818 |
+
"SICKFr": 80.2,
|
| 14819 |
+
"STS12": 82.53,
|
| 14820 |
+
"STS13": 88.05,
|
| 14821 |
+
"STS14": 84.83,
|
| 14822 |
+
"STS15": 91.02,
|
| 14823 |
+
"STS16": 87.32,
|
| 14824 |
+
"STS17 (ar-ar)": 82.71,
|
| 14825 |
+
"STS17 (nl-en)": 87.44,
|
| 14826 |
+
"STS17 (fr-en)": 86.28,
|
| 14827 |
+
"STS17 (ko-ko)": 84.31,
|
| 14828 |
+
"STS17 (es-es)": 89.4,
|
| 14829 |
+
"STS17 (en-ar)": 81.83,
|
| 14830 |
+
"STS17 (en-de)": 87.61,
|
| 14831 |
+
"STS17 (en-tr)": 80.47,
|
| 14832 |
+
"STS17 (es-en)": 87.03,
|
| 14833 |
+
"STS17 (en-en)": 90.33,
|
| 14834 |
+
"STS17 (it-en)": 88.25,
|
| 14835 |
+
"STS22 (es-en)": 78.65,
|
| 14836 |
+
"STS22 (de)": 61.53,
|
| 14837 |
+
"STS22 (zh)": 67.4,
|
| 14838 |
+
"STS22 (es)": 68.45,
|
| 14839 |
+
"STS22 (de-fr)": 65.52,
|
| 14840 |
+
"STS22 (pl)": 40.97,
|
| 14841 |
+
"STS22 (ru)": 65.17,
|
| 14842 |
+
"STS22 (en)": 68.67,
|
| 14843 |
+
"STS22 (fr)": 82.25,
|
| 14844 |
+
"STS22 (es-it)": 75.25,
|
| 14845 |
+
"STS22 (pl-en)": 74.86,
|
| 14846 |
+
"STS22 (fr-pl)": 61.98,
|
| 14847 |
+
"STS22 (ar)": 59.52,
|
| 14848 |
+
"STS22 (it)": 79.45,
|
| 14849 |
+
"STS22 (tr)": 67.2,
|
| 14850 |
+
"STS22 (zh-en)": 71.45,
|
| 14851 |
+
"STS22 (de-pl)": 54.6,
|
| 14852 |
+
"STS22 (de-en)": 58.22,
|
| 14853 |
+
"STSB": 81.68,
|
| 14854 |
+
"STSBenchmark": 88.39,
|
| 14855 |
+
"STSBenchmarkMultilingualSTS (de)": 85.38,
|
| 14856 |
+
"STSBenchmarkMultilingualSTS (nl)": 84.29,
|
| 14857 |
+
"STSBenchmarkMultilingualSTS (pl)": 83.42,
|
| 14858 |
+
"STSBenchmarkMultilingualSTS (fr)": 84.94,
|
| 14859 |
+
"STSBenchmarkMultilingualSTS (pt)": 84.9,
|
| 14860 |
+
"STSBenchmarkMultilingualSTS (it)": 84.38,
|
| 14861 |
+
"STSBenchmarkMultilingualSTS (es)": 86.15,
|
| 14862 |
+
"STSBenchmarkMultilingualSTS (zh)": 82.54,
|
| 14863 |
+
"STSBenchmarkMultilingualSTS (en)": 88.39,
|
| 14864 |
+
"STSBenchmarkMultilingualSTS (ru)": 83.86
|
| 14865 |
+
}
|
| 14866 |
+
]
|
| 14867 |
+
},
|
| 14868 |
+
"Summarization": {
|
| 14869 |
+
"cosine_spearman": [
|
| 14870 |
+
{
|
| 14871 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14872 |
+
"SummEval": 30.46,
|
| 14873 |
+
"SummEvalFr": 31.72
|
| 14874 |
+
},
|
| 14875 |
+
{
|
| 14876 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14877 |
+
"SummEval": 30.46,
|
| 14878 |
+
"SummEvalFr": 31.72
|
| 14879 |
+
}
|
| 14880 |
+
]
|
| 14881 |
+
},
|
| 14882 |
+
"MultilabelClassification": {
|
| 14883 |
+
"accuracy": [
|
| 14884 |
+
{
|
| 14885 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14886 |
+
"CEDRClassification": 50.01,
|
| 14887 |
+
"SensitiveTopicsClassification": 32.29
|
| 14888 |
+
}
|
| 14889 |
+
]
|
| 14890 |
+
},
|
| 14891 |
+
"InstructionRetrieval": {
|
| 14892 |
+
"p-MRR": [
|
| 14893 |
+
{
|
| 14894 |
+
"Model": "<a target=\"_blank\" style=\"text-decoration: underline\" href=\"https://huggingface.co/intfloat/multilingual-e5-large-instruct\">multilingual-e5-large-instruct</a>",
|
| 14895 |
+
"Core17InstructionRetrieval": 1.82,
|
| 14896 |
+
"News21InstructionRetrieval": 1.5,
|
| 14897 |
+
"Robust04InstructionRetrieval": -4.52
|
| 14898 |
+
}
|
| 14899 |
+
]
|
| 14900 |
+
}
|
| 14901 |
+
},
|
| 14902 |
"multilingual-e5-small": {
|
| 14903 |
"BitextMining": {
|
| 14904 |
"f1": [
|
model_meta.yaml
CHANGED
|
@@ -1072,6 +1072,15 @@ model_meta:
|
|
| 1072 |
is_proprietary: false
|
| 1073 |
is_sentence_transformers_compatible: true
|
| 1074 |
uses_instruct: false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1075 |
multilingual-e5-small:
|
| 1076 |
link: https://huggingface.co/intfloat/multilingual-e5-small
|
| 1077 |
seq_len: 512
|
|
|
|
| 1072 |
is_proprietary: false
|
| 1073 |
is_sentence_transformers_compatible: true
|
| 1074 |
uses_instruct: false
|
| 1075 |
+
multilingual-e5-large-instruct:
|
| 1076 |
+
link: https://huggingface.co/intfloat/multilingual-e5-large-instruct
|
| 1077 |
+
seq_len: 514
|
| 1078 |
+
size: 560
|
| 1079 |
+
dim: 1024
|
| 1080 |
+
is_external: true
|
| 1081 |
+
is_proprietary: false
|
| 1082 |
+
is_sentence_transformers_compatible: true
|
| 1083 |
+
uses_instruct: true
|
| 1084 |
multilingual-e5-small:
|
| 1085 |
link: https://huggingface.co/intfloat/multilingual-e5-small
|
| 1086 |
seq_len: 512
|