Tarek07 commited on
Commit
9c99408
·
verified ·
1 Parent(s): 1340300

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -13,7 +13,7 @@ tags:
13
  license: llama3.3
14
  ---
15
  After having some fair success with merging some of my favorite Llama models, particularly on
16
- Tarek07/Progenitor-V1.1-LLaMa-70B, which was based on the excellent Steelskull/L3.3-MS-Nevoria-70b merge, just with a couple extra ingredients and different merge methods. I then decided to try my hand at working the new deepseek-ai/DeepSeek-R1-Distill-Llama-70B in. I have decided to try mix it in 3 different ways. For this first model I decided to use Deepseek as the base, using the same formula as my Progenitor V1.1 model. From early testing it seems very promising.
17
  # merge
18
 
19
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
13
  license: llama3.3
14
  ---
15
  After having some fair success with merging some of my favorite Llama models, particularly on
16
+ Tarek07/Progenitor-V1.1-LLaMa-70B, which was based on the excellent Steelskull/L3.3-MS-Nevoria-70b merge, just with a couple extra ingredients and different merge methods. I then decided to try my hand at working the new deepseek-ai/DeepSeek-R1-Distill-Llama-70B in. I have decided to try mix it in 3 different ways. For this first model I decided to use Deepseek as the base, using the same formula as my Progenitor V1.1 model. From early testing it seems very promising. But I think Delta beat this one.
17
  # merge
18
 
19
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).