Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ tags:
|
|
12 |
- merge
|
13 |
license: llama3.3
|
14 |
---
|
15 |
-
My Thalassic series includes my favorite Llama 3 models with Deepseek R1, using various methods I made 4 merges
|
16 |
# merge
|
17 |
|
18 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
|
|
12 |
- merge
|
13 |
license: llama3.3
|
14 |
---
|
15 |
+
My Thalassic series includes my favorite Llama 3 models with Deepseek R1, using various methods I made 4 merges. Thalassic Delta, which I made with an SCE merge, was the best of the set. I actually got some advice from Steelskull about how I had my top k parameter way too high. So I decided to lower it for this merge. Now because I feel Delta was still a decent model with the top k on 1, I halved the top k parameter to 0.50 which is basically the max recommended setting.
|
16 |
# merge
|
17 |
|
18 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|