p1atdev commited on
Commit
587f3dd
·
verified ·
1 Parent(s): 8f2b9f8

Upload rope_migration.yml

Browse files
Files changed (1) hide show
  1. rope-15/rope_migration.yml +99 -0
rope-15/rope_migration.yml ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
3
+ pretrained_model_name_or_path: fal/AuraFlow-v0.3
4
+
5
+ dtype: bfloat16
6
+
7
+ denoiser:
8
+ use_flash_attn: true
9
+
10
+ use_rope: True
11
+ rope_theta: 10000
12
+ rope_dim_sizes: [32, 112, 112]
13
+
14
+ noise_prediction_loss: true
15
+ migration_loss: true
16
+ prior_preservation_loss: false
17
+ migration_freezing_threshold: 1.0e-7
18
+
19
+ peft:
20
+ type: lora
21
+ rank: 4
22
+ alpha: 1.0
23
+ dropout: 0.0
24
+
25
+ dtype: bfloat16
26
+
27
+ include_keys:
28
+ - ".mlp."
29
+ - ".attn."
30
+
31
+ exclude_keys:
32
+ - "text_encoder"
33
+ - "vae"
34
+ - "t_embedder"
35
+ - "final_linear"
36
+ - regex: .*\.mod[CX]{1,2} # exclude modulation layers (modC, modCX, modX)
37
+
38
+ dataset:
39
+ folder: "datasets/pexels-50k"
40
+ num_repeats: 1
41
+ batch_size: 2
42
+
43
+ bucket_base_size: 1024
44
+ step: 128
45
+ min_size: 384
46
+ do_upscale: false
47
+
48
+ caption_processors: []
49
+
50
+ optimizer:
51
+ name: "schedulefree.RAdamScheduleFree"
52
+ args:
53
+ lr: 0.05
54
+
55
+ tracker:
56
+ project_name: "auraflow-rope-1"
57
+ loggers:
58
+ - wandb
59
+
60
+ saving:
61
+ strategy:
62
+ per_epochs: 1
63
+ per_steps: 500
64
+ # save_last: true
65
+
66
+ callbacks:
67
+ - type: "hf_hub" # or "hf_hub" to push to hub
68
+ name: "rope-15"
69
+ save_dir: "./output/rope-15"
70
+
71
+ hub_id: "p1atdev/afv03-lora"
72
+ dir_in_repo: "rope-15"
73
+
74
+ preview:
75
+ strategy:
76
+ per_epochs: 1
77
+ per_steps: 100
78
+
79
+ callbacks:
80
+ - type: "discord"
81
+ url: "masked"
82
+
83
+ data:
84
+ path: "./projects/rope/preview.yml"
85
+
86
+ seed: 42
87
+ num_train_epochs: 10
88
+
89
+ trainer:
90
+ # debug_mode: "1step"
91
+
92
+ gradient_checkpointing: true
93
+ gradient_accumulation_steps: 8
94
+
95
+ torch_compile: true
96
+ torch_compile_args:
97
+ mode: max-autotune
98
+ fullgraph: true
99
+ fp32_matmul_precision: "medium"