p1atdev commited on
Commit
f896df0
·
verified ·
1 Parent(s): 55d5a82

Upload rope_migration.yml

Browse files
Files changed (1) hide show
  1. rope-14/rope_migration.yml +103 -0
rope-14/rope_migration.yml ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
3
+ pretrained_model_name_or_path: fal/AuraFlow-v0.3
4
+
5
+ dtype: bfloat16
6
+
7
+ denoiser:
8
+ use_flash_attn: true
9
+
10
+ use_rope: True
11
+ rope_theta: 10000
12
+ rope_dim_sizes: [32, 112, 112]
13
+
14
+ noise_prediction_loss: true
15
+ migration_loss: true
16
+ prior_preservation_loss: false
17
+ migration_round_threshold: 1.0e-7
18
+
19
+ peft:
20
+ type: lora
21
+ rank: 4
22
+ alpha: 1.0
23
+ dropout: 0.0
24
+
25
+ dtype: bfloat16
26
+
27
+ include_keys:
28
+ - ".mlp."
29
+ - ".attn."
30
+
31
+ exclude_keys:
32
+ - "text_encoder"
33
+ - "vae"
34
+ - "t_embedder"
35
+ - "final_linear"
36
+ - regex: .*\.mod[CX]{1,2} # exclude modulation layers (modC, modCX, modX)
37
+
38
+ dataset:
39
+ folder: "data/pexels-1k-random"
40
+ num_repeats: 2
41
+ batch_size: 2
42
+
43
+ bucket_base_size: 1024
44
+ step: 128
45
+ min_size: 384
46
+ do_upscale: false
47
+
48
+ caption_processors: []
49
+
50
+ optimizer:
51
+ name: "schedulefree.RAdamScheduleFree"
52
+ # name: "bitsandbytes.optim.AdamW8bit"
53
+ args:
54
+ lr: 0.005
55
+
56
+ scheduler:
57
+ # name: "torch.optim.lr_scheduler.ConstantLR"
58
+ # args: {}
59
+
60
+ tracker:
61
+ project_name: "auraflow-rope-1"
62
+ loggers:
63
+ - wandb
64
+
65
+ saving:
66
+ strategy:
67
+ per_epochs: 1
68
+ per_steps: null
69
+ save_last: true
70
+
71
+ callbacks:
72
+ - type: "hf_hub" # or "hf_hub" to push to hub
73
+ name: "rope-14"
74
+ save_dir: "./output/rope-14"
75
+
76
+ hub_id: "p1atdev/afv03-lora"
77
+ dir_in_repo: "rope-14"
78
+
79
+ preview:
80
+ strategy:
81
+ per_epochs: 1
82
+ per_steps: 100
83
+
84
+ callbacks:
85
+ - type: "discord"
86
+ url: "https://discord.com/api/webhooks/1327274657308610560/3T8Jj1DqkF7_T0eiYWThHu9dvqlcSYXPFFrGq8Dsd-k7lgp7MYyF5WN7KS65SW4K-v1l"
87
+
88
+ data:
89
+ path: "./projects/rope/preview.yml"
90
+
91
+ seed: 42
92
+ num_train_epochs: 20
93
+
94
+ trainer:
95
+ # debug_mode: "1step"
96
+
97
+ gradient_checkpointing: true
98
+
99
+ torch_compile: true
100
+ torch_compile_args:
101
+ mode: max-autotune
102
+ fullgraph: true
103
+ fp32_matmul_precision: "medium"