RichardErkhov commited on
Commit
d3984a9
·
verified ·
1 Parent(s): c037335

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +89 -0
README.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ mergekit-ties-xzdpqzs - AWQ
11
+ - Model creator: https://huggingface.co/Hjgugugjhuhjggg/
12
+ - Original model: https://huggingface.co/Hjgugugjhuhjggg/mergekit-ties-xzdpqzs/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ base_model:
20
+ - huyhoangt2201/llama-3.2-1b-sql_finetuned_billingual_3.0_merged
21
+ - jayavibhav/llama3.2_1b_CoT
22
+ - huyhoangt2201/llama-3.2-1b-chat-sql3-merged
23
+ - autoprogrammer/Llama-3.2-1B-Instruct-MGSM8K-sft1
24
+ - Alelcv27/llama3.2-1b-math-code
25
+ - unsloth/Llama-3.2-1B-Instruct-bnb-4bit
26
+ - meta-llama/Llama-3.2-1B
27
+ - student-abdullah/Llama3.2-1B_Hinglish-Medicine-Dataset_Finetuning_28-09
28
+ - meta-llama/Llama-3.2-1B-Instruct
29
+ - MLking2/llama-3.2-1b-medical
30
+ - autoprogrammer/Llama-3.2-1B-Instruct-medmcqa-zh-linear
31
+ library_name: transformers
32
+ tags:
33
+ - mergekit
34
+ - merge
35
+
36
+ ---
37
+ # merge
38
+
39
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
40
+
41
+ ## Merge Details
42
+ ### Merge Method
43
+
44
+ This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) as a base.
45
+
46
+ ### Models Merged
47
+
48
+ The following models were included in the merge:
49
+ * [huyhoangt2201/llama-3.2-1b-sql_finetuned_billingual_3.0_merged](https://huggingface.co/huyhoangt2201/llama-3.2-1b-sql_finetuned_billingual_3.0_merged)
50
+ * [jayavibhav/llama3.2_1b_CoT](https://huggingface.co/jayavibhav/llama3.2_1b_CoT)
51
+ * [huyhoangt2201/llama-3.2-1b-chat-sql3-merged](https://huggingface.co/huyhoangt2201/llama-3.2-1b-chat-sql3-merged)
52
+ * [autoprogrammer/Llama-3.2-1B-Instruct-MGSM8K-sft1](https://huggingface.co/autoprogrammer/Llama-3.2-1B-Instruct-MGSM8K-sft1)
53
+ * [Alelcv27/llama3.2-1b-math-code](https://huggingface.co/Alelcv27/llama3.2-1b-math-code)
54
+ * [unsloth/Llama-3.2-1B-Instruct-bnb-4bit](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct-bnb-4bit)
55
+ * [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B)
56
+ * [student-abdullah/Llama3.2-1B_Hinglish-Medicine-Dataset_Finetuning_28-09](https://huggingface.co/student-abdullah/Llama3.2-1B_Hinglish-Medicine-Dataset_Finetuning_28-09)
57
+ * [MLking2/llama-3.2-1b-medical](https://huggingface.co/MLking2/llama-3.2-1b-medical)
58
+ * [autoprogrammer/Llama-3.2-1B-Instruct-medmcqa-zh-linear](https://huggingface.co/autoprogrammer/Llama-3.2-1B-Instruct-medmcqa-zh-linear)
59
+
60
+ ### Configuration
61
+
62
+ The following YAML configuration was used to produce this model:
63
+
64
+ ```yaml
65
+ merge_method: ties
66
+ architectures: ["transformer"]
67
+ base_model: meta-llama/Llama-3.2-1B-Instruct
68
+
69
+ models:
70
+ - model: Alelcv27/llama3.2-1b-math-code
71
+ - model: huyhoangt2201/llama-3.2-1b-sql_finetuned_billingual_3.0_merged
72
+ - model: autoprogrammer/Llama-3.2-1B-Instruct-MGSM8K-sft1
73
+ - model: meta-llama/Llama-3.2-1B-Instruct
74
+ - model: autoprogrammer/Llama-3.2-1B-Instruct-medmcqa-zh-linear
75
+ - model: meta-llama/Llama-3.2-1B
76
+ - model: unsloth/Llama-3.2-1B-Instruct-bnb-4bit
77
+ - model: MLking2/llama-3.2-1b-medical
78
+ - model: jayavibhav/llama3.2_1b_CoT
79
+ - model: huyhoangt2201/llama-3.2-1b-chat-sql3-merged
80
+ - model: student-abdullah/Llama3.2-1B_Hinglish-Medicine-Dataset_Finetuning_28-09
81
+
82
+ parameters:
83
+ density: 0.5
84
+ weight: 1.0
85
+ int8_mask: true
86
+ normalize: true
87
+ ```
88
+
89
+