removed gemma-9b
Browse files
README.md
CHANGED
@@ -40,7 +40,7 @@ The model delivers best-in-class performance across Indic tasks and a promising
|
|
40 |
- Matches or exceeds performance of models much larger (x6) on multilingual Indic generation tasks including creative writing, summarization, and translation;
|
41 |
- Stronger Indian cultural context relevance - scored the highest in manual evaluation with multiple models in an anonymised setting;
|
42 |
- Delivers top-3 performance on 5 (out of 7) tasks in BharatBench among much larger open source and commercial models.
|
43 |
-
- Available in
|
44 |
|
45 |
## Model Developer
|
46 |
- OLA Krutrim Team
|
@@ -110,15 +110,15 @@ We use the LM Evaluation Harness to evaluate our model on the En benchmarks task
|
|
110 |
### BharatBench
|
111 |
The existing Indic benchmarks are not natively in Indian languages, rather, they are translations of existing En benchmarks. They do not sufficiently capture the linguistic nuances of Indian languages and aspects of Indian culture. Towards that Krutrim released BharatBench - a natively Indic benchmark that encompasses the linguistic and cultural diversity of the Indic region, ensuring that the evaluations are relevant and representative of real-world use cases in India.
|
112 |
|
113 |
-
| Benchmark | Metric | Krutrim-1-7B | MN-12B-Instruct | Krutrim-2-12B | llama-3.1-8B-Instruct | llama-3.1-70B-Instruct | Gemma-2-
|
114 |
-
|
115 |
-
| Indian Cultural Context (0-shot) | Bert Score | 0.86 | 0.56 | 0.88 | 0.87 | 0.88 | 0.87 | 0.
|
116 |
-
| Grammar Correction (5-shot) | Bert Score | 0.96 | 0.94 | 0.98 | 0.95 | 0.98 | 0.96 | 0.
|
117 |
-
| Multi Turn (0-shot) | Bert Score | 0.88 | 0.87 | 0.91 | 0.88 | 0.90 | 0.89 | 0.
|
118 |
-
| Multi Turn Comprehension (0-shot) | Bert Score | 0.90 | 0.89 | 0.92 | 0.92 | 0.93 | 0.91 | 0.
|
119 |
-
| Multi Turn Translation (0-shot) | Bert Score | 0.85 | 0.87 | 0.92 | 0.89 | 0.91 | 0.
|
120 |
-
| Text Classification (5-shot) | Accuracy | 0.61 | 0.71 | 0.76 | 0.72 | 0.88 | 0.
|
121 |
-
| Named Entity Recognition (5-shot) | Accuracy | 0.31 | 0.51 | 0.53 | 0.55 | 0.61 | 0.
|
122 |
|
123 |
### Qualitative Results
|
124 |
Below are the results from manual evaluation of prompt-response pairs across languages and task categories. Scores are between 1-5 (higher the better). Model names were anonymised during the evaluation.
|
|
|
40 |
- Matches or exceeds performance of models much larger (x6) on multilingual Indic generation tasks including creative writing, summarization, and translation;
|
41 |
- Stronger Indian cultural context relevance - scored the highest in manual evaluation with multiple models in an anonymised setting;
|
42 |
- Delivers top-3 performance on 5 (out of 7) tasks in BharatBench among much larger open source and commercial models.
|
43 |
+
- Available in instruction-tuned version
|
44 |
|
45 |
## Model Developer
|
46 |
- OLA Krutrim Team
|
|
|
110 |
### BharatBench
|
111 |
The existing Indic benchmarks are not natively in Indian languages, rather, they are translations of existing En benchmarks. They do not sufficiently capture the linguistic nuances of Indian languages and aspects of Indian culture. Towards that Krutrim released BharatBench - a natively Indic benchmark that encompasses the linguistic and cultural diversity of the Indic region, ensuring that the evaluations are relevant and representative of real-world use cases in India.
|
112 |
|
113 |
+
| Benchmark | Metric | Krutrim-1-7B | MN-12B-Instruct | Krutrim-2-12B | llama-3.1-8B-Instruct | llama-3.1-70B-Instruct | Gemma-2-27B-Instruct | GPT-4o |
|
114 |
+
|-------------------------------------|------------|--------------|-----------------|---------------|------------------------|------------------------|---------------------|--------|
|
115 |
+
| Indian Cultural Context (0-shot) | Bert Score | 0.86 | 0.56 | 0.88 | 0.87 | 0.88 | 0.87 | 0.89 |
|
116 |
+
| Grammar Correction (5-shot) | Bert Score | 0.96 | 0.94 | 0.98 | 0.95 | 0.98 | 0.96 | 0.97 |
|
117 |
+
| Multi Turn (0-shot) | Bert Score | 0.88 | 0.87 | 0.91 | 0.88 | 0.90 | 0.89 | 0.92 |
|
118 |
+
| Multi Turn Comprehension (0-shot) | Bert Score | 0.90 | 0.89 | 0.92 | 0.92 | 0.93 | 0.91 | 0.94 |
|
119 |
+
| Multi Turn Translation (0-shot) | Bert Score | 0.85 | 0.87 | 0.92 | 0.89 | 0.91 | 0.91 | 0.92 |
|
120 |
+
| Text Classification (5-shot) | Accuracy | 0.61 | 0.71 | 0.76 | 0.72 | 0.88 | 0.86 | 0.89 |
|
121 |
+
| Named Entity Recognition (5-shot) | Accuracy | 0.31 | 0.51 | 0.53 | 0.55 | 0.61 | 0.65 | 0.65 |
|
122 |
|
123 |
### Qualitative Results
|
124 |
Below are the results from manual evaluation of prompt-response pairs across languages and task categories. Scores are between 1-5 (higher the better). Model names were anonymised during the evaluation.
|