p1atdev commited on
Commit
4c72058
·
verified ·
1 Parent(s): a80d8fa

Upload rope_migration.yml

Browse files
Files changed (1) hide show
  1. rope-20/rope_migration.yml +98 -0
rope-20/rope_migration.yml ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
3
+ pretrained_model_name_or_path: fal/AuraFlow-v0.3
4
+
5
+ dtype: bfloat16
6
+
7
+ denoiser:
8
+ use_flash_attn: true
9
+
10
+ use_rope: True
11
+ rope_theta: 10000
12
+ rope_dim_sizes: [32, 112, 112]
13
+
14
+ noise_prediction_loss: true
15
+ migration_loss: false
16
+ prior_preservation_loss: false
17
+
18
+ peft:
19
+ type: lora
20
+ rank: 4
21
+ alpha: 1.0
22
+ dropout: 0.0
23
+
24
+ dtype: bfloat16
25
+
26
+ include_keys:
27
+ - ".mlp."
28
+ - ".attn."
29
+
30
+ exclude_keys:
31
+ - "text_encoder"
32
+ - "vae"
33
+ - "t_embedder"
34
+ - "final_linear"
35
+ - regex: .*\.mod[CX]{1,2} # exclude modulation layers (modC, modCX, modX)
36
+
37
+ dataset:
38
+ folder: "masked"
39
+ num_repeats: 1
40
+ batch_size: 2
41
+
42
+ bucket_base_size: 1024
43
+ step: 128
44
+ min_size: 384
45
+ do_upscale: false
46
+
47
+ caption_processors: []
48
+
49
+ optimizer:
50
+ name: "schedulefree.RAdamScheduleFree"
51
+ args:
52
+ lr: 0.03
53
+
54
+ tracker:
55
+ project_name: "auraflow-rope-1"
56
+ loggers:
57
+ - wandb
58
+
59
+ saving:
60
+ strategy:
61
+ per_epochs: 1
62
+ per_steps: 500
63
+ # save_last: true
64
+
65
+ callbacks:
66
+ - type: "hf_hub" # or "hf_hub" to push to hub
67
+ name: "rope-20"
68
+ save_dir: "./output/rope-20"
69
+
70
+ hub_id: "p1atdev/afv03-lora"
71
+ dir_in_repo: "rope-20"
72
+
73
+ preview:
74
+ strategy:
75
+ per_epochs: 1
76
+ per_steps: 100
77
+
78
+ callbacks:
79
+ - type: "discord"
80
+ url: "masked"
81
+
82
+ data:
83
+ path: "./projects/rope/preview.yml"
84
+
85
+ seed: 42
86
+ num_train_epochs: 10
87
+
88
+ trainer:
89
+ # debug_mode: "1step"
90
+
91
+ gradient_checkpointing: true
92
+ gradient_accumulation_steps: 8
93
+
94
+ torch_compile: true
95
+ torch_compile_args:
96
+ mode: max-autotune
97
+ fullgraph: true
98
+ fp32_matmul_precision: "medium"