Tarek07 commited on
Commit
b742cf3
·
verified ·
1 Parent(s): f041e4e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -12,7 +12,7 @@ tags:
12
  - merge
13
  license: llama3.3
14
  ---
15
- My Thalassic series includes my favorite Llama 3 models with Deepseek R1, using various methods I made 4 merges, of which Thalassic Delta, which I made with an SCE merge, was the best of the set. I actually got some advice from Steelskull about how I had my top k parameter way too high. So I decided to lower it for this merge. Now because I feel Delta was still a decent model with the top k on 1, I halved the top k parameter to 0.50 which is basically the max recommended setting.
16
  # merge
17
 
18
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
12
  - merge
13
  license: llama3.3
14
  ---
15
+ My Thalassic series includes my favorite Llama 3 models with Deepseek R1, using various methods I made 4 merges. Thalassic Delta, which I made with an SCE merge, was the best of the set. I actually got some advice from Steelskull about how I had my top k parameter way too high. So I decided to lower it for this merge. Now because I feel Delta was still a decent model with the top k on 1, I halved the top k parameter to 0.50 which is basically the max recommended setting.
16
  # merge
17
 
18
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).