Update README.md
Browse files
README.md
CHANGED
@@ -12,9 +12,9 @@ tags:
|
|
12 |
|
13 |
# Model Card for Phoenix
|
14 |
|
15 |
-
**Phoenix** is a model trained using Direct Preference Optimization (DPO)
|
16 |
-
In contrast to zephyr and notus this model has been trained using german instruction and dpo
|
17 |
-
and HuggingFaceH4/ultrafeedback_binarized
|
18 |
While the mistral model performs really well, it is not really suitable for the german language. Therefore we have used the fantastic LeoLM/leo-mistral-hessianai-7b.
|
19 |
Thanks to the new type of training, Phoenix is not only able to compete with the Mistral model from LeoLM but also **beats the Llama-70b-chat model in 2 mt-bench categories**.
|
20 |
This model **wouldn't have been possible without the amazing work of Huggingface, LeoLM, openbnb, Argilla the Alma-Team and many others of the AI community**.
|
|
|
12 |
|
13 |
# Model Card for Phoenix
|
14 |
|
15 |
+
**Phoenix** is a model trained using Direct Preference Optimization (DPO) for the german language. Its training procedure follows the process of the alignment-handbook from Huggingface.
|
16 |
+
In contrast to zephyr and notus this model has been trained using german instruction and dpo data. In detail, a german translation of HuggingFaceH4/ultrachat_200k
|
17 |
+
and HuggingFaceH4/ultrafeedback_binarized were created in addition to a series of allready available instruction datasets. The LLM haoranxu/ALMA-13B was used for this.
|
18 |
While the mistral model performs really well, it is not really suitable for the german language. Therefore we have used the fantastic LeoLM/leo-mistral-hessianai-7b.
|
19 |
Thanks to the new type of training, Phoenix is not only able to compete with the Mistral model from LeoLM but also **beats the Llama-70b-chat model in 2 mt-bench categories**.
|
20 |
This model **wouldn't have been possible without the amazing work of Huggingface, LeoLM, openbnb, Argilla the Alma-Team and many others of the AI community**.
|