Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: sacla
|
4 |
+
license_link: >-
|
5 |
+
https://huggingface.co/stabilityai/stable-diffusion-3.5-large/blob/main/LICENSE.md
|
6 |
+
base_model:
|
7 |
+
- stabilityai/stable-diffusion-3.5-large
|
8 |
+
base_model_relation: quantized
|
9 |
+
---
|
10 |
+
## Overview
|
11 |
+
These models are made to work with [stable-diffusion.cpp](https://github.com/leejet/stable-diffusion.cpp) release [master-ac54e00](https://github.com/leejet/stable-diffusion.cpp/releases/tag/master-ac54e00) onwards. Support for other inference backends is not guarenteed.
|
12 |
+
|
13 |
+
Quantized using this PR https://github.com/leejet/stable-diffusion.cpp/pull/447
|
14 |
+
|
15 |
+
Normal K-quants are not working properly with SD3.5-Large models because around 90% of the weights are in tensors whose shape doesn't match the 256 superblock size of K-quants and therefore can't be quantized this way. Mixing quantization types allows us to take adventage of the better fidelity of k-quants to some extent while keeping the model file size relatively small.
|
16 |
+
|
17 |
+
## Files:
|
18 |
+
|
19 |
+
### Mixed Types:
|
20 |
+
|
21 |
+
TODO
|
22 |
+
|
23 |
+
### Legacy types:
|
24 |
+
|
25 |
+
TODO
|
26 |
+
|
27 |
+
## Outputs:
|
28 |
+
|
29 |
+
Sorted by model size (Note that q4_0 and q4_k_4_0 are the exact same size)
|
30 |
+
|
31 |
+
| Quantization | Robot girl | Text | Cute kitten |
|
32 |
+
| ------------------ | -------------------------------- | ---------------------------------- | ---------------------------------- |
|
33 |
+
|
34 |
+
|
35 |
+
Generated with a modified version of sdcpp with [this PR](https://github.com/leejet/stable-diffusion.cpp/pull/397) applied to enable clip timestep embeddings support.
|
36 |
+
|
37 |
+
Text encoders used: q4_k quant of t5xxl, full precision clip_g, and q8 quant of [ViT-L-14-TEXT-detail-improved-hiT-GmP-TE-only-HF](https://huggingface.co/zer0int/CLIP-GmP-ViT-L-14) in place of clip_l.
|
38 |
+
|
39 |
+
Full prompts and settings in png metadata.
|