Update README.md
Browse files
README.md
CHANGED
@@ -23,9 +23,9 @@ The models in this merge are some of my favorites and I found I liked all of the
|
|
23 |
|
24 |
Model_stock was the method used, it's very straightforward and quite fast, the bottleneck seemed to be my NVMe drive.
|
25 |
|
26 |
-
All source models use ChatML prompt formatting and it responds very well.
|
27 |
|
28 |
-
|
29 |
|
30 |
## Merge Details
|
31 |
### Merge Method
|
|
|
23 |
|
24 |
Model_stock was the method used, it's very straightforward and quite fast, the bottleneck seemed to be my NVMe drive.
|
25 |
|
26 |
+
All source models use ChatML prompt formatting and it responds very well. Consider the following settings (thanks Geechan!): Temp 1.25, MinP 0.02, XTC 0.15/probability 0.5, DRY 0.8. All other samplers neutralized. Chuluun seems to be able to work with higher temperatures than other Qwen models without losing coherency.
|
27 |
|
28 |
+
Konnect has released their [Qwenception](https://huggingface.co/Konnect1221/The-Inception-Presets-Methception-LLamaception-Qwenception) sysprompts and settings, which work quite well with Chuluun.
|
29 |
|
30 |
## Merge Details
|
31 |
### Merge Method
|