Unconditional Image Generation
English
encoreus commited on
Commit
a1e5393
·
verified ·
1 Parent(s): aa154c6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -3
README.md CHANGED
@@ -19,11 +19,13 @@ pipeline_tag: unconditional-image-generation
19
 
20
  **Transformer AutoRegressive Flow Model**
21
 
22
- The TarFlow is proposed by [Zhai et al., 2024], which introduces stacks of autoregressive Transformer blocks (similar to MAF) into the building of affine coupling layers to do Non-Volume Preserving, combined with guidance and denoising }, finally achieves state-of-the-art results across multiple benchmarks.
 
23
 
24
- It's sampling process is extremely slow, and we want to accelerate it in [Liu and Qin, 2025]. In experiments, we find that the model parameters are not available in [Zhai et al., 2024], so we retrain TarFlow models and upload them.
 
25
 
26
- As metioned in [Zhai et al., 2024], a TarFlow model can be denoted as P-Ch-T-K-pε, with
27
  patch size (P), model channel size (Ch), number of autoregressive flow blocks (T), the number of attention layers in each flow (K), the best input noise variance pε that yields the best sampling quality for generation tasks.
28
 
29
  We trained five models:
@@ -38,12 +40,24 @@ We trained five models:
38
 
39
  - ImageNet (64x64) conditioanl: imagenet_model_4_1024_8_8_0.05.pth
40
 
 
 
 
 
 
 
 
 
 
41
 
42
  The sampling traces maybe look like this:
43
 
44
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/682459b20ee49a8c3822a525/bx0U7vnZSu9SFoV2gtERI.png)
45
  From top to bottom: Img128cond, Img64cond (patch4), Img64uncond, AFHQ. From left to right: noise, Block 7-0, denoised image.
46
 
 
 
 
47
  [1] Zhai S, Zhang R, Nakkiran P, et al. Normalizing flows are capable generative models[J]. arXiv preprint arXiv:2412.06329, 2024.
48
 
49
  [2] Liu, B. \& Qin, Z. Accelerate TarFlow Sampling with GS-Jacobi Iteration. (2025), https://arxiv.org/abs/2505.12849
 
19
 
20
  **Transformer AutoRegressive Flow Model**
21
 
22
+ The TarFlow is proposed by [[Zhai et al., 2024]](http://arxiv.org/abs/2412.06329), which introduces stacks of autoregressive Transformer blocks (similar to MAF) into the building of affine coupling layers to do Non-Volume Preserving, combined with guidance and denoising, finally achieves state-of-the-art results across multiple benchmarks.
23
+ Associated code can be found at https://github.com/apple/ml-tarflow.
24
 
25
+ It's sampling process is extremely slow, and we want to accelerate it in [[Liu and Qin, 2025]](https://arxiv.org/abs/2505.12849), code in https://github.com/encoreus/GS-Jacobi_for_TarFlow.
26
+ In experiments, we find that the model parameters are not available in original paper, so we retrain TarFlow models and upload them.
27
 
28
+ As metioned in [[Zhai et al., 2024]](http://arxiv.org/abs/2412.06329), a TarFlow model can be denoted as P-Ch-T-K-pε, with
29
  patch size (P), model channel size (Ch), number of autoregressive flow blocks (T), the number of attention layers in each flow (K), the best input noise variance pε that yields the best sampling quality for generation tasks.
30
 
31
  We trained five models:
 
40
 
41
  - ImageNet (64x64) conditioanl: imagenet_model_4_1024_8_8_0.05.pth
42
 
43
+ We also compute the stats for the true data distribution which can be used to calculate FID.
44
+
45
+ - AFHQ (256x256) conditional: afhq_256_fid_stats.pth
46
+
47
+ - ImageNet (128x128) conditional: imagenet_128_fid_stats.pth
48
+
49
+ - ImageNet (64x64) unconditional: imagenet64_64_fid_stats.pth
50
+
51
+ - ImageNet (64x64) conditional: imagenet_64_fid_stats.pth
52
 
53
  The sampling traces maybe look like this:
54
 
55
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/682459b20ee49a8c3822a525/bx0U7vnZSu9SFoV2gtERI.png)
56
  From top to bottom: Img128cond, Img64cond (patch4), Img64uncond, AFHQ. From left to right: noise, Block 7-0, denoised image.
57
 
58
+
59
+
60
+
61
  [1] Zhai S, Zhang R, Nakkiran P, et al. Normalizing flows are capable generative models[J]. arXiv preprint arXiv:2412.06329, 2024.
62
 
63
  [2] Liu, B. \& Qin, Z. Accelerate TarFlow Sampling with GS-Jacobi Iteration. (2025), https://arxiv.org/abs/2505.12849