metadata
base_model:
- datatab/Yugo55-GPT-v4
- datatab/Yugo55-GPT-DPO-v1-chkp-600
library_name: transformers
tags:
- mergekit
- merge
- text-generation-inference
- transformers
- mistral
license: mit
language:
- sr
datasets:
- datatab/alpaca-cleaned-serbian-full
- datatab/ultrafeedback_binarized
- datatab/open-orca-slim-serbian
Yugo60-GPT
- Developed by: datatab
- License: mit
🏆 Results
Results obtained through the Serbian LLM evaluation, released by Aleksa Gordić: serbian-llm-eval
- Evaluation was conducted on a 4-bit version of the model due to hardware resource constraints.
MODEL | ARC-E | ARC-C | Hellaswag | BoolQ | Winogrande | OpenbookQA | PiQA |
---|---|---|---|---|---|---|---|
*Yugo55-GPT-v4-4bit | 51.41 | 36.00 | 57.51 | 80.92 | 65.75 | 34.70 | 70.54 |
Yugo55A-GPT | 51.52 | 37.78 | 57.52 | 84.40 | 65.43 | 35.60 | 69.43 |
Yugo60-GPT | tbd | tbd | tbd | tbd | tbd | tbd | tbd |