Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,17 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# FLAN-T5-XXL Fused Model
|
2 |
+
|
3 |
+
This repository contains a fused version of FLAN-T5-XXL, combining the split files available at [Google's FLAN-T5-XXL repository](https://huggingface.co/google/flan-t5-xxl). These files have been merged for easier use in AI applications, including image generation models.
|
4 |
+
|
5 |
+
Additionally, a GGUF-optimized version of this model is available at [dumb-dev's Hugging Face repository](https://huggingface.co/dumb-dev/flan-t5-xxl-gguf). This version is tailored for use in environments optimized for GGUF format, such as certain LLM inference frameworks.
|
6 |
+
|
7 |
+
## Files
|
8 |
+
- `flan_t5_xxl_fp32.safetensors`: Full-precision FP32 model.
|
9 |
+
- `flan_t5_xxl_fp16.safetensors`: Half-precision FP16 model for memory-efficient inference.
|
10 |
+
|
11 |
+
## License
|
12 |
+
This model is provided under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0).
|
13 |
+
The uploader does not claim any rights over the model.
|
14 |
+
|
15 |
+
## Acknowledgments
|
16 |
+
- Original source: [Google's FLAN-T5-XXL repository](https://huggingface.co/google/flan-t5-xxl).
|
17 |
+
- GGUF version: [dumb-dev's Hugging Face repository](https://huggingface.co/dumb-dev/flan-t5-xxl-gguf).
|