Update README.md
Browse files
README.md
CHANGED
@@ -5,14 +5,11 @@ license: apache-2.0
|
|
5 |
* Quantization of Qwen2.5 14B for edge devices 7.3Gb footprint
|
6 |
* One of the best models I tried in Spanish.
|
7 |
* Original model: https://huggingface.co/djuna/Q2.5-Veltha-14B-0.5
|
8 |
-
* Models Merged
|
9 |
-
*
|
10 |
-
|
11 |
-
*
|
12 |
-
*
|
13 |
-
* EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
|
14 |
-
* v000000/Qwen2.5-Lumen-14B
|
15 |
-
`
|
16 |
|
17 |
* All quants made using imatrix option with dataset from here
|
18 |
* Using llama.cpp compiled with CUDA support for quantization and inference:
|
|
|
5 |
* Quantization of Qwen2.5 14B for edge devices 7.3Gb footprint
|
6 |
* One of the best models I tried in Spanish.
|
7 |
* Original model: https://huggingface.co/djuna/Q2.5-Veltha-14B-0.5
|
8 |
+
* Models Merged:
|
9 |
+
* huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2
|
10 |
+
* allura-org/TQ2.5-14B-Aletheia-v1
|
11 |
+
* EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
|
12 |
+
* v000000/Qwen2.5-Lumen-14B
|
|
|
|
|
|
|
13 |
|
14 |
* All quants made using imatrix option with dataset from here
|
15 |
* Using llama.cpp compiled with CUDA support for quantization and inference:
|