Update README.md
Browse files
README.md
CHANGED
@@ -12,9 +12,13 @@ tags:
|
|
12 |
|
13 |
# Model Card for Phoenix
|
14 |
|
15 |
-
|
16 |
-
|
17 |
-
|
|
|
|
|
|
|
|
|
18 |
|
19 |
## MT-Bench-DE Scores
|
20 |
Phoenix beats the LeoLM-Mistral model in all categories except for coding and humanities.
|
|
|
12 |
|
13 |
# Model Card for Phoenix
|
14 |
|
15 |
+
**Phoenix** is a model trained using Direct Preference Optimization (DPO). This model is the first version, which has been trained following the process of the alignment-handbook from Huggingface.
|
16 |
+
In contrast to zephyr and notus this model has been trained using german instruction and dpo-data. In detail, a german translation of HuggingFaceH4/ultrachat_200k
|
17 |
+
and HuggingFaceH4/ultrafeedback_binarized was created in addition to a series of instruction datasets. The LLM haoranxu/ALMA-13B was used for this.
|
18 |
+
While the mistral model performs really well, it is not really suitable for the german language. Therefore we have used the fantastic LeoLM/leo-mistral-hessianai-7b.
|
19 |
+
Thanks to the new type of training, Phoenix is not only able to compete with the Mistral model from LeoLM but also **beats the Llama-70b-chat model in 2 mt-bench categories**.
|
20 |
+
This model **wouldn't have been possible without the amazing work of Huggingface, LeoLM, openbnb, Argilla the Alma-Team and many others of the AI community**.
|
21 |
+
i would like to personally thank all AI researchers who make the training of such models possible
|
22 |
|
23 |
## MT-Bench-DE Scores
|
24 |
Phoenix beats the LeoLM-Mistral model in all categories except for coding and humanities.
|