PyTorch
mistral
Krutrim
language-model
krutrim-admin commited on
Commit
fb9bab0
·
verified ·
1 Parent(s): 0cc38b7

removed gemma-9b

Browse files
Files changed (1) hide show
  1. README.md +10 -10
README.md CHANGED
@@ -40,7 +40,7 @@ The model delivers best-in-class performance across Indic tasks and a promising
40
  - Matches or exceeds performance of models much larger (x6) on multilingual Indic generation tasks including creative writing, summarization, and translation;
41
  - Stronger Indian cultural context relevance - scored the highest in manual evaluation with multiple models in an anonymised setting;
42
  - Delivers top-3 performance on 5 (out of 7) tasks in BharatBench among much larger open source and commercial models.
43
- - Available in both pre-trained and instruction-tuned versions
44
 
45
  ## Model Developer
46
  - OLA Krutrim Team
@@ -110,15 +110,15 @@ We use the LM Evaluation Harness to evaluate our model on the En benchmarks task
110
  ### BharatBench
111
  The existing Indic benchmarks are not natively in Indian languages, rather, they are translations of existing En benchmarks. They do not sufficiently capture the linguistic nuances of Indian languages and aspects of Indian culture. Towards that Krutrim released BharatBench - a natively Indic benchmark that encompasses the linguistic and cultural diversity of the Indic region, ensuring that the evaluations are relevant and representative of real-world use cases in India.
112
 
113
- | Benchmark | Metric | Krutrim-1-7B | MN-12B-Instruct | Krutrim-2-12B | llama-3.1-8B-Instruct | llama-3.1-70B-Instruct | Gemma-2-9B-Instruct | Gemma-2-27B-Instruct | GPT-4o |
114
- |-------------------------------------|------------|--------------|-----------------|---------------|------------------------|------------------------|---------------------|---------------------|--------|
115
- | Indian Cultural Context (0-shot) | Bert Score | 0.86 | 0.56 | 0.88 | 0.87 | 0.88 | 0.87 | 0.87 | 0.89 |
116
- | Grammar Correction (5-shot) | Bert Score | 0.96 | 0.94 | 0.98 | 0.95 | 0.98 | 0.96 | 0.96 | 0.97 |
117
- | Multi Turn (0-shot) | Bert Score | 0.88 | 0.87 | 0.91 | 0.88 | 0.90 | 0.89 | 0.89 | 0.92 |
118
- | Multi Turn Comprehension (0-shot) | Bert Score | 0.90 | 0.89 | 0.92 | 0.92 | 0.93 | 0.91 | 0.91 | 0.94 |
119
- | Multi Turn Translation (0-shot) | Bert Score | 0.85 | 0.87 | 0.92 | 0.89 | 0.91 | 0.90 | 0.91 | 0.92 |
120
- | Text Classification (5-shot) | Accuracy | 0.61 | 0.71 | 0.76 | 0.72 | 0.88 | 0.82 | 0.86 | 0.89 |
121
- | Named Entity Recognition (5-shot) | Accuracy | 0.31 | 0.51 | 0.53 | 0.55 | 0.61 | 0.61 | 0.65 | 0.65 |
122
 
123
  ### Qualitative Results
124
  Below are the results from manual evaluation of prompt-response pairs across languages and task categories. Scores are between 1-5 (higher the better). Model names were anonymised during the evaluation.
 
40
  - Matches or exceeds performance of models much larger (x6) on multilingual Indic generation tasks including creative writing, summarization, and translation;
41
  - Stronger Indian cultural context relevance - scored the highest in manual evaluation with multiple models in an anonymised setting;
42
  - Delivers top-3 performance on 5 (out of 7) tasks in BharatBench among much larger open source and commercial models.
43
+ - Available in instruction-tuned version
44
 
45
  ## Model Developer
46
  - OLA Krutrim Team
 
110
  ### BharatBench
111
  The existing Indic benchmarks are not natively in Indian languages, rather, they are translations of existing En benchmarks. They do not sufficiently capture the linguistic nuances of Indian languages and aspects of Indian culture. Towards that Krutrim released BharatBench - a natively Indic benchmark that encompasses the linguistic and cultural diversity of the Indic region, ensuring that the evaluations are relevant and representative of real-world use cases in India.
112
 
113
+ | Benchmark | Metric | Krutrim-1-7B | MN-12B-Instruct | Krutrim-2-12B | llama-3.1-8B-Instruct | llama-3.1-70B-Instruct | Gemma-2-27B-Instruct | GPT-4o |
114
+ |-------------------------------------|------------|--------------|-----------------|---------------|------------------------|------------------------|---------------------|--------|
115
+ | Indian Cultural Context (0-shot) | Bert Score | 0.86 | 0.56 | 0.88 | 0.87 | 0.88 | 0.87 | 0.89 |
116
+ | Grammar Correction (5-shot) | Bert Score | 0.96 | 0.94 | 0.98 | 0.95 | 0.98 | 0.96 | 0.97 |
117
+ | Multi Turn (0-shot) | Bert Score | 0.88 | 0.87 | 0.91 | 0.88 | 0.90 | 0.89 | 0.92 |
118
+ | Multi Turn Comprehension (0-shot) | Bert Score | 0.90 | 0.89 | 0.92 | 0.92 | 0.93 | 0.91 | 0.94 |
119
+ | Multi Turn Translation (0-shot) | Bert Score | 0.85 | 0.87 | 0.92 | 0.89 | 0.91 | 0.91 | 0.92 |
120
+ | Text Classification (5-shot) | Accuracy | 0.61 | 0.71 | 0.76 | 0.72 | 0.88 | 0.86 | 0.89 |
121
+ | Named Entity Recognition (5-shot) | Accuracy | 0.31 | 0.51 | 0.53 | 0.55 | 0.61 | 0.65 | 0.65 |
122
 
123
  ### Qualitative Results
124
  Below are the results from manual evaluation of prompt-response pairs across languages and task categories. Scores are between 1-5 (higher the better). Model names were anonymised during the evaluation.