Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
23.7
TFLOPS
8
16
382
Mike Afton
ifmain
Follow
Mi6paulino's profile picture
shtefcs's profile picture
Lulalaby's profile picture
11 followers
·
40 following
ifmain
AI & ML interests
None yet
Recent Activity
liked
a model
about 2 hours ago
fal/Qwen-Image-Edit-2511-Multiple-Angles-LoRA
reacted
to
MonsterMMORPG
's
post
with 👀
about 2 hours ago
Compared Quality and Speed Difference (with CUDA 13 & Sage Attention) of BF16 vs GGUF Q8 vs FP8 Scaled vs NVFP4 for Z Image Turbo, FLUX Dev, FLUX SRPO, FLUX Kontext, FLUX 2 - Full 4K step by step tutorial also published Full 4K tutorial : https://youtu.be/XDzspWgnzxI Check above full 4K tutorial to learn more and see uncompressed original quality and size images It was always wondered how much quality and speed difference exists between BF16, GGUF, FP8 Scaled and NVFP4 precisions. In this tutorial I have compared all these precision and quantization variants for both speed and quality. The results are pretty surprising. Moreover, we have developed and published NVFP4 model quant generator app and FP8 Scaled quant generator apps. The links of the apps are below if you want to use them. Furthermore, upgrading ComfyUI to CUDA 13 with properly compiled libraries is now very much recommended. We have observed some noticeable performance gains with CUDA 13. So for both SwarmUI and ComfyUI solo users, CUDA 13 ComfyUI is now recommended.
reacted
to
MonsterMMORPG
's
post
with 👀
about 2 hours ago
Compared Quality and Speed Difference (with CUDA 13 & Sage Attention) of BF16 vs GGUF Q8 vs FP8 Scaled vs NVFP4 for Z Image Turbo, FLUX Dev, FLUX SRPO, FLUX Kontext, FLUX 2 - Full 4K step by step tutorial also published Full 4K tutorial : https://youtu.be/XDzspWgnzxI Check above full 4K tutorial to learn more and see uncompressed original quality and size images It was always wondered how much quality and speed difference exists between BF16, GGUF, FP8 Scaled and NVFP4 precisions. In this tutorial I have compared all these precision and quantization variants for both speed and quality. The results are pretty surprising. Moreover, we have developed and published NVFP4 model quant generator app and FP8 Scaled quant generator apps. The links of the apps are below if you want to use them. Furthermore, upgrading ComfyUI to CUDA 13 with properly compiled libraries is now very much recommended. We have observed some noticeable performance gains with CUDA 13. So for both SwarmUI and ComfyUI solo users, CUDA 13 ComfyUI is now recommended.
View all activity
Organizations
ifmain
's datasets
5
Sort: Recently updated
ifmain/search_in_text-01
Viewer
•
Updated
Nov 14, 2024
•
2k
•
5
•
1
ifmain/text-moderation-02-multilingual
Viewer
•
Updated
Oct 13, 2024
•
1.55M
•
19
•
1
ifmain/comment-translation-01
Viewer
•
Updated
Oct 13, 2024
•
1.46M
•
23
•
1
ifmain/text-moderation-02-large
Viewer
•
Updated
Jun 27, 2024
•
410k
•
96
•
8
ifmain/text-moderation-01
Viewer
•
Updated
Feb 4, 2024
•
70k
•
33
•
8