KaraKaraWitch commited on
Commit
5b9de41
·
verified ·
1 Parent(s): eb6dc4e

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +80 -0
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Steelskull/L3-MS-Astoria-70b
4
+ - Sao10K/L3-70B-Euryale-v2.1
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+
10
+ ---
11
+
12
+ # Eurstoria-120B
13
+
14
+ ![](Unholy.png)
15
+
16
+ A Llama-3 Decoder only model by combining 2x finetuned Llama-3 70B models into 1.
17
+
18
+ ## Merge Details
19
+ ### Merge Method
20
+
21
+ This model was merged using the passthrough merge method.
22
+
23
+ ### Models Merged
24
+
25
+ The following models were included in the merge:
26
+ * [Steelskull/L3-MS-Astoria-70b](https://huggingface.co/Steelskull/L3-MS-Astoria-70b)
27
+ * [Sao10K/L3-70B-Euryale-v2.1](https://huggingface.co/Sao10K/L3-70B-Euryale-v2.1)
28
+
29
+ ### Configuration
30
+
31
+ The following YAML configuration was used to produce this model:
32
+
33
+ ```yaml
34
+ slices:
35
+ - sources:
36
+ - model: Steelskull/L3-MS-Astoria-70b
37
+ layer_range: [0, 16]
38
+ - sources:
39
+ - model: Sao10K/L3-70B-Euryale-v2.1
40
+ layer_range: [8, 24]
41
+ - sources:
42
+ - model: Steelskull/L3-MS-Astoria-70b
43
+ layer_range: [25, 40]
44
+ - sources:
45
+ - model: Sao10K/L3-70B-Euryale-v2.1
46
+ layer_range: [33, 48]
47
+ - sources:
48
+ - model: Steelskull/L3-MS-Astoria-70b
49
+ layer_range: [41, 56]
50
+ - sources:
51
+ - model: Sao10K/L3-70B-Euryale-v2.1
52
+ layer_range: [49, 64]
53
+ - sources:
54
+ - model: Steelskull/L3-MS-Astoria-70b
55
+ layer_range: [57, 72]
56
+ - sources:
57
+ - model: Sao10K/L3-70B-Euryale-v2.1
58
+ layer_range: [65, 80]
59
+
60
+ merge_method: passthrough
61
+ dtype: float16
62
+ ```
63
+
64
+ # Prompting Format
65
+
66
+ As both models are based off from instruct, Use the default [Llama 3 Instruct format](https://huggingface.co/Sao10K/L3-70B-Euryale-v2.1/blob/main/Euryale-v2.1-Llama-3-Instruct.json) would be fine.
67
+
68
+ # Screenshots
69
+
70
+ None. Unfortunately or Fortunately.
71
+
72
+ # Benchmarks
73
+
74
+ No one is willing to bench a 120B... right?
75
+
76
+ # Acknowledgements
77
+
78
+ The original ratios come from Goliath and are used as is. As such thanks to both [Alpindale](https://huggingface.co/alpindale) and [@Undi95](https://huggingface.co/Undi95) for their inital work.
79
+
80
+ Credits goes to [@chargoddard](https://huggingface.co/chargoddard) for developing the framework used to merge the model - [mergekit](https://github.com/cg123/mergekit).