Update README.md
Browse files
README.md
CHANGED
|
@@ -78,11 +78,11 @@ For reference, Warner et al.'s ModernBERT uses 1.72T tokens for stage 1, 250B to
|
|
| 78 |
JSTS, JNLI, and JCoLA from [JGLUE](https://aclanthology.org/2022.lrec-1.317/) were used.
|
| 79 |
Evaluation code can be found at https://github.com/speed1313/bert-eval
|
| 80 |
|
| 81 |
-
| Model | JSTS (pearson) | JNLI (accuracy) | JCoLA(accuracy) | Avg |
|
| 82 |
|-------------------------------------------------------|--------|--------|---------|--------------|
|
| 83 |
| tohoku-nlp/bert-base-japanese-v3 | 0.920 | 0.912 | 0.880 | 0.904 |
|
| 84 |
| sbintuitions/modernbert-ja-130m | 0.916 | 0.927 | 0.868 | 0.904 |
|
| 85 |
-
| sbintuitions/modernbert-ja-310m | 0.932 | 0.933 | 0.883 | 0.916 |
|
| 86 |
-
| speed/llm-jp-modernbert-base-v4-ja-stage2-200k | 0.918 | 0.913 | 0.844 | 0.892 |
|
| 87 |
|
| 88 |
|
|
|
|
| 78 |
JSTS, JNLI, and JCoLA from [JGLUE](https://aclanthology.org/2022.lrec-1.317/) were used.
|
| 79 |
Evaluation code can be found at https://github.com/speed1313/bert-eval
|
| 80 |
|
| 81 |
+
| Model | JSTS (pearson) | JNLI (accuracy) | JCoLA (accuracy) | Avg |
|
| 82 |
|-------------------------------------------------------|--------|--------|---------|--------------|
|
| 83 |
| tohoku-nlp/bert-base-japanese-v3 | 0.920 | 0.912 | 0.880 | 0.904 |
|
| 84 |
| sbintuitions/modernbert-ja-130m | 0.916 | 0.927 | 0.868 | 0.904 |
|
| 85 |
+
| sbintuitions/modernbert-ja-310m | **0.932** | **0.933** | **0.883** | **0.916** |
|
| 86 |
+
| **speed/llm-jp-modernbert-base-v4-ja-stage2-200k** | 0.918 | 0.913 | 0.844 | 0.892 |
|
| 87 |
|
| 88 |
|