DRXD1000 commited on
Commit
3570ad2
·
1 Parent(s): f07212d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -12,9 +12,13 @@ tags:
12
 
13
  # Model Card for Phoenix
14
 
15
- Phoenixis a model trained using Direct Preference Optimization (DPO). This model is the first version, which has been trained following the process of the Alignment-Handbook from Huggingface. In contrast to zephyr and notus this model has been trained using german instruction and dpo-data. In detail, a German translation of HuggingFaceH4/ultrachat_200k
16
- and HuggingFaceH4/ultrafeedback_binarized was created in addition to a series of instruction datasets. The LLM haoranxu/ALMA-13B was used for this. While the mistral model performs really well, it is not really suitable for the german language. Therefore we have used the fantastic LeoLM/leo-mistral-hessianai-7b.
17
- Thanks to the new type of training, Phoenix is not only able to compete with the Mistral model from LeoLM but also **beats the Llama-70b-chat model in 2 mt-bench categories**This model **wouldn't have been possible without the amazing work of Huggingface, LeoLM, openbnb, Argilla the Alma-Team and many others of the AI community**
 
 
 
 
18
 
19
  ## MT-Bench-DE Scores
20
  Phoenix beats the LeoLM-Mistral model in all categories except for coding and humanities.
 
12
 
13
  # Model Card for Phoenix
14
 
15
+ **Phoenix** is a model trained using Direct Preference Optimization (DPO). This model is the first version, which has been trained following the process of the alignment-handbook from Huggingface.
16
+ In contrast to zephyr and notus this model has been trained using german instruction and dpo-data. In detail, a german translation of HuggingFaceH4/ultrachat_200k
17
+ and HuggingFaceH4/ultrafeedback_binarized was created in addition to a series of instruction datasets. The LLM haoranxu/ALMA-13B was used for this.
18
+ While the mistral model performs really well, it is not really suitable for the german language. Therefore we have used the fantastic LeoLM/leo-mistral-hessianai-7b.
19
+ Thanks to the new type of training, Phoenix is not only able to compete with the Mistral model from LeoLM but also **beats the Llama-70b-chat model in 2 mt-bench categories**.
20
+ This model **wouldn't have been possible without the amazing work of Huggingface, LeoLM, openbnb, Argilla the Alma-Team and many others of the AI community**.
21
+ i would like to personally thank all AI researchers who make the training of such models possible
22
 
23
  ## MT-Bench-DE Scores
24
  Phoenix beats the LeoLM-Mistral model in all categories except for coding and humanities.