Blazgo commited on
Commit
29eee5a
·
verified ·
1 Parent(s): f198c1e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -37
README.md CHANGED
@@ -1,40 +1,7 @@
1
  ---
2
  base_model:
3
- - Sakalti/Saka-14B
4
- - mkurman/Qwen2.5-14B-DeepSeek-R1-1M
5
  library_name: transformers
6
- tags:
7
- - mergekit
8
- - merge
9
-
10
- ---
11
- # merge
12
-
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ## Merge Details
16
- ### Merge Method
17
-
18
- This model was merged using the [SLERP](https://en.wikipedia.org/wiki/Slerp) merge method.
19
-
20
- ### Models Merged
21
-
22
- The following models were included in the merge:
23
- * [Sakalti/Saka-14B](https://huggingface.co/Sakalti/Saka-14B)
24
- * [mkurman/Qwen2.5-14B-DeepSeek-R1-1M](https://huggingface.co/mkurman/Qwen2.5-14B-DeepSeek-R1-1M)
25
-
26
- ### Configuration
27
-
28
- The following YAML configuration was used to produce this model:
29
-
30
- ```yaml
31
- models:
32
- - model: Sakalti/Saka-14B
33
- - model: mkurman/Qwen2.5-14B-DeepSeek-R1-1M
34
- merge_method: slerp
35
- base_model: mkurman/Qwen2.5-14B-DeepSeek-R1-1M
36
- dtype: bfloat16
37
- parameters:
38
- t: [1, 0.5, 0, 0.5, 1] # A shaped curve: Hermes for input & output, WizardMath in the middle layers
39
-
40
- ```
 
1
  ---
2
  base_model:
3
+ - Qwen/Qwen2.5-14B-Instruct-1M
4
+ - deepseek-ai/DeepSeek-R1
5
  library_name: transformers
6
+ license: cc-by-nc-nd-4.0
7
+ ---