Text Generation
Transformers
Safetensors
Serbian
mistral
mergekit
Merge
text-generation-inference
conversational
Yugo60-GPT / README.md
datatab's picture
Update README.md
da0da5c verified
|
raw
history blame
1.79 kB
metadata
base_model:
  - datatab/Yugo55-GPT-v4
  - datatab/Yugo55-GPT-DPO-v1-chkp-600
library_name: transformers
tags:
  - mergekit
  - merge
  - text-generation-inference
  - transformers
  - mistral
license: mit
language:
  - sr
datasets:
  - datatab/alpaca-cleaned-serbian-full
  - datatab/ultrafeedback_binarized
  - datatab/open-orca-slim-serbian

Yugo60-GPT

  • Developed by: datatab
  • License: mit

🏆 Results

Results obtained through the Serbian LLM evaluation, released by Aleksa Gordić: serbian-llm-eval

  • Evaluation was conducted on a 4-bit version of the model due to hardware resource constraints.
MODEL ARC-E ARC-C Hellaswag BoolQ Winogrande OpenbookQA PiQA
*Yugo55-GPT-v4-4bit 51.41 36.00 57.51 80.92 65.75 34.70 70.54
Yugo55A-GPT 51.52 37.78 57.52 84.40 65.43 35.60 69.43
Yugo60-GPT tbd tbd tbd tbd tbd tbd tbd