rk91 commited on
Commit
26a9974
·
1 Parent(s): 8309f4e

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +328 -0
app.py ADDED
@@ -0,0 +1,328 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ """DreamBooth_Stable_Diffusion_V2.ipynb
3
+
4
+ Automatically generated by Colaboratory.
5
+
6
+ Original file is located at
7
+ https://colab.research.google.com/github/KaliYuga-ai/DreamBoothV2fork/blob/main/DreamBooth_Stable_Diffusion_V2.ipynb
8
+
9
+ ##DreamBooth with Stable Diffusion V2
10
+
11
+ This notebook is [KaliYuga](https://twitter.com/KaliYuga_ai)'s very basic fork of [Shivam Shrirao](https://github.com/ShivamShrirao)'s DreamBooth notebook. In addition to a vew minor formatting and QoL additions, I've added Stable Diffusion V2 as the default training option and optimized the training settings to reflect what I've found to be the best general ones. They are only suggestions; feel free to tweak anything and everything if my defaults don't do it for you.
12
+
13
+ **I also [wrote a guide](https://peakd.com/hive-158694/@kaliyuga/training-a-dreambooth-model-using-stable-diffusion-v2-and-very-little-code)** that should take you through building a dataset and training a model using this notebook. If this is your first time creating a model from scratch, I reccommend you check it out!
14
+ """
15
+
16
+ #@markdown Check type of GPU and VRAM available.
17
+ !nvidia-smi --query-gpu=name,memory.total,memory.free --format=csv,noheader
18
+
19
+ """https://github.com/KaliYuga-ai/diffusers/tree/main/examples/dreambooth
20
+
21
+ ## Install Requirements
22
+ """
23
+
24
+ # Commented out IPython magic to ensure Python compatibility.
25
+ !wget -q https://github.com/ShivamShrirao/diffusers/raw/main/examples/dreambooth/train_dreambooth.py
26
+ !wget -q https://github.com/ShivamShrirao/diffusers/raw/main/scripts/convert_diffusers_to_original_stable_diffusion.py
27
+ # %pip install -qq git+https://github.com/ShivamShrirao/diffusers
28
+ # %pip install -q -U --pre triton
29
+ # %pip install -q accelerate==0.12.0 transformers ftfy bitsandbytes gradio natsort
30
+
31
+ #@title Login to HuggingFace 🤗
32
+
33
+ #@markdown You need to accept the model license before downloading or using the Stable Diffusion weights. Please, visit the [model card](https://huggingface.co/stabilityai/stable-diffusion-2), read the license and tick the checkbox if you agree. You have to be a registered user in 🤗 Hugging Face Hub, and you'll also need to use an access token for the code to work.
34
+ # https://huggingface.co/settings/tokens
35
+ !mkdir -p ~/.huggingface
36
+ HUGGINGFACE_TOKEN = "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" #@param {type:"string"}
37
+ !echo -n "{HUGGINGFACE_TOKEN}" > ~/.huggingface/token
38
+
39
+ """### Install xformers from precompiled wheel."""
40
+
41
+ # Commented out IPython magic to ensure Python compatibility.
42
+ # %pip install -q https://github.com/metrolobo/xformers_wheels/releases/download/1d31a3ac_various_6/xformers-0.0.14.dev0-cp37-cp37m-linux_x86_64.whl
43
+ # These were compiled on Tesla T4, should also work on P100, thanks to https://github.com/metrolobo
44
+
45
+ # If precompiled wheels don't work, install it with the following command. It will take around 40 minutes to compile.
46
+ # %pip install git+https://github.com/facebookresearch/xformers@1d31a3a#egg=xformers
47
+
48
+ """## Settings and run"""
49
+
50
+ #@markdown If model weights should be saved directly in google drive (takes around 4-5 GB).
51
+ save_to_gdrive = True #@param {type:"boolean"}
52
+ if save_to_gdrive:
53
+ from google.colab import drive
54
+ drive.mount('/content/drive')
55
+
56
+ #@markdown Name/Path of the initial model.
57
+ MODEL_NAME = "stabilityai/stable-diffusion-2" #@param {type:"string"}
58
+
59
+ #@markdown Enter the directory name to save model at.
60
+
61
+ OUTPUT_DIR = "stable_diffusion_weights/ps1theme" #@param {type:"string"}
62
+ if save_to_gdrive:
63
+ OUTPUT_DIR = "/content/drive/MyDrive/" + OUTPUT_DIR
64
+ else:
65
+ OUTPUT_DIR = "/content/" + OUTPUT_DIR
66
+
67
+ print(f"[*] Weights will be saved at {OUTPUT_DIR}")
68
+
69
+ !mkdir -p $OUTPUT_DIR
70
+
71
+ """### Start Training
72
+
73
+ Use the table below to choose the best flags based on your memory and speed requirements. Tested on Tesla T4 GPU.
74
+
75
+
76
+ | `fp16` | `train_batch_size` | `gradient_accumulation_steps` | `gradient_checkpointing` | `use_8bit_adam` | GB VRAM usage | Speed (it/s) |
77
+ | ---- | ------------------ | ----------------------------- | ----------------------- | --------------- | ---------- | ------------ |
78
+ | fp16 | 1 | 1 | TRUE | TRUE | 9.92 | 0.93 |
79
+ | no | 1 | 1 | TRUE | TRUE | 10.08 | 0.42 |
80
+ | fp16 | 2 | 1 | TRUE | TRUE | 10.4 | 0.66 |
81
+ | fp16 | 1 | 1 | FALSE | TRUE | 11.17 | 1.14 |
82
+ | no | 1 | 1 | FALSE | TRUE | 11.17 | 0.49 |
83
+ | fp16 | 1 | 2 | TRUE | TRUE | 11.56 | 1 |
84
+ | fp16 | 2 | 1 | FALSE | TRUE | 13.67 | 0.82 |
85
+ | fp16 | 1 | 2 | FALSE | TRUE | 13.7 | 0.83 |
86
+ | fp16 | 1 | 1 | TRUE | FALSE | 15.79 | 0.77 |
87
+ ------------------------------------------------------------------------------
88
+
89
+
90
+ - `--gradient_checkpointing` flag is enabled by default; it reduces VRAM usage to 9.92 GB usage.
91
+
92
+ - remove `--use_8bit_adam` flag for full precision. Requires 15.79 GB with `--gradient_checkpointing` else 17.8 GB.
93
+
94
+ - remove `--train_text_encoder` flag to reduce memory usage further, degrades output quality. NOT RECCOMMENDED.
95
+
96
+ ### Define Your Concepts List
97
+ You can add multiple concepts here. Try tweaking `--max_train_steps` accordingly.
98
+ It's a good idea to test class prompts in Stable Diffusion V2 before committing to them. If the images V2 generates at a CFG of 7 and 50 steps aren't great, consider a different class prompt.
99
+ """
100
+
101
+ concepts_list = [
102
+ {
103
+ "instance_prompt": "ps1theme",
104
+ "class_prompt": "women",
105
+ "instance_data_dir": "/content/drive/MyDrive/ps1theme",
106
+ "class_data_dir": "/content/data/women"
107
+ },
108
+
109
+ # {
110
+ # "instance_prompt": "photo of zwx dog",
111
+ # "class_prompt": "photo of a dog",
112
+ # "instance_data_dir": "/content/data/zwx",
113
+ # "class_data_dir": "/content/data/dog"
114
+ # },
115
+ # {
116
+ # "instance_prompt": "photo of ukj person",
117
+ # "class_prompt": "photo of a person",
118
+ # "instance_data_dir": "/content/data/ukj",
119
+ # "class_data_dir": "/content/data/person"
120
+ # }
121
+ ]
122
+
123
+ # `class_data_dir` contains regularization images
124
+ import json
125
+ import os
126
+ for c in concepts_list:
127
+ os.makedirs(c["instance_data_dir"], exist_ok=True)
128
+
129
+ with open("concepts_list.json", "w") as f:
130
+ json.dump(concepts_list, f, indent=4)
131
+
132
+ """### Image Upload"""
133
+
134
+ #@markdown Upload your images by running this cell.
135
+
136
+ #@markdown OR
137
+
138
+ #@markdown Alteranately, add your dataset to google drive and then copy its path into `instance_data_dir"`, above. You can also use the file manager on the left panel to upload (drag and drop) to each `instance_data_dir`; it uploads faster than running the cell.
139
+
140
+ import os
141
+ from google.colab import files
142
+ import shutil
143
+
144
+ for c in concepts_list:
145
+ print(f"Uploading instance images for `{c['instance_prompt']}`")
146
+ uploaded = files.upload()
147
+ for filename in uploaded.keys():
148
+ dst_path = os.path.join(c['instance_data_dir'], filename)
149
+ shutil.move(filename, dst_path)
150
+
151
+ """### Training Settings
152
+ The Learning Rate in this notebook has been sped up from the default LR in previous Dreambooth notebooks; training runs slower on SD V 2. This might not be the best LR for all usecases, but does well for all the datasets I (KaliYuga) have tried so far.
153
+ Please note, `gradient_checkpointing` is enabled by default. I think it produces better results, and it reduces VRAM.
154
+ """
155
+
156
+ !accelerate launch train_dreambooth.py \
157
+ --pretrained_model_name_or_path=$MODEL_NAME \
158
+ --pretrained_vae_name_or_path="stabilityai/sd-vae-ft-mse" \
159
+ --output_dir=$OUTPUT_DIR \
160
+ --revision="fp16" \
161
+ --with_prior_preservation --prior_loss_weight=1.0 \
162
+ --seed=1337 \
163
+ --resolution=512 \
164
+ --train_batch_size=1 \
165
+ --train_text_encoder \
166
+ --mixed_precision="fp16" \
167
+ --use_8bit_adam \
168
+ --gradient_accumulation_steps=1 \
169
+ --gradient_checkpointing \
170
+ --learning_rate=4e-6 \
171
+ --lr_scheduler="constant" \
172
+ --lr_warmup_steps=0 \
173
+ --num_class_images=50 \
174
+ --sample_batch_size=4 \
175
+ --max_train_steps=5000 \
176
+ --save_interval=500 \
177
+ --save_sample_prompt="ps1theme" \
178
+ --concepts_list="concepts_list.json"
179
+
180
+ # Reduce the `--save_interval` to lower than `--max_train_steps` to save weights from intermediate steps.
181
+ # `--save_sample_prompt` can be same as `--instance_prompt` to generate intermediate samples (saved along with weights in samples directory).
182
+
183
+ """### Testing your new model
184
+
185
+ Once your model has finished training (or has reached a checkpoint you like), run the following cells to test it out.
186
+ """
187
+
188
+ #@markdown Specify the weights directory to use (leave blank for latest)
189
+ WEIGHTS_DIR = "/content/drive/MyDrive/stable_diffusion_weights/ps1theme/2000" #@param {type:"string"}
190
+ if WEIGHTS_DIR == "":
191
+ from natsort import natsorted
192
+ from glob import glob
193
+ import os
194
+ WEIGHTS_DIR = natsorted(glob(OUTPUT_DIR + os.sep + "*"))[-1]
195
+ print(f"[*] WEIGHTS_DIR={WEIGHTS_DIR}")
196
+
197
+ from transformers import BertForMaskedLM
198
+
199
+ model = BertForMaskedLM.from_pretrained("remi/bertabs-finetuned-extractive-abstractive-summarization")
200
+
201
+ from google.colab import drive
202
+ drive.mount('/content/drive')
203
+
204
+ #@markdown Run to generate a grid of preview images from the last saved weights.
205
+ import os
206
+ import matplotlib.pyplot as plt
207
+ import matplotlib.image as mpimg
208
+
209
+ weights_folder = OUTPUT_DIR
210
+ folders = sorted([f for f in os.listdir(weights_folder) if f != "0"], key=lambda x: int(x))
211
+
212
+ row = len(folders)
213
+ col = len(os.listdir(os.path.join(weights_folder, folders[0], "samples")))
214
+ scale = 4
215
+ fig, axes = plt.subplots(row, col, figsize=(col*scale, row*scale), gridspec_kw={'hspace': 0, 'wspace': 0})
216
+
217
+ for i, folder in enumerate(folders):
218
+ folder_path = os.path.join(weights_folder, folder)
219
+ image_folder = os.path.join(folder_path, "samples")
220
+ images = [f for f in os.listdir(image_folder)]
221
+ for j, image in enumerate(images):
222
+ if row == 1:
223
+ currAxes = axes[j]
224
+ else:
225
+ currAxes = axes[i, j]
226
+ if i == 0:
227
+ currAxes.set_title(f"Image {j}")
228
+ if j == 0:
229
+ currAxes.text(-0.1, 0.5, folder, rotation=0, va='center', ha='center', transform=currAxes.transAxes)
230
+ image_path = os.path.join(image_folder, image)
231
+ img = mpimg.imread(image_path)
232
+ currAxes.imshow(img, cmap='gray')
233
+ currAxes.axis('off')
234
+
235
+ plt.tight_layout()
236
+ plt.savefig('grid.png', dpi=72)
237
+
238
+ """#### Convert weights to ckpt to use in web UIs like AUTOMATIC1111."""
239
+
240
+ #@markdown Run conversion.
241
+ ckpt_path = WEIGHTS_DIR + "/model.ckpt"
242
+
243
+ half_arg = ""
244
+ #@markdown Whether to convert to fp16, takes half the space (2GB).
245
+ fp16 = True #@param {type: "boolean"}
246
+ if fp16:
247
+ half_arg = "--half"
248
+ !python convert_diffusers_to_original_stable_diffusion.py --model_path $WEIGHTS_DIR --checkpoint_path $ckpt_path $half_arg
249
+ print(f"[*] Converted ckpt saved at {ckpt_path}")
250
+
251
+ """#### Inference"""
252
+
253
+ from transformers import CLIPProcessor, CLIPModel,CLIPTokenizer
254
+
255
+ import torch
256
+ from torch import autocast
257
+ from diffusers import StableDiffusionPipeline, DDIMScheduler
258
+ from IPython.display import display
259
+
260
+ model_path = WEIGHTS_DIR # If you want to use previously trained model saved in gdrive, replace this with the full path of model in gdrive
261
+
262
+ scheduler = DDIMScheduler(beta_start=0.00085, beta_end=0.012, beta_schedule="scaled_linear", clip_sample=False, set_alpha_to_one=False)
263
+ pipe = StableDiffusionPipeline.from_pretrained(model_path, scheduler=scheduler, safety_checker=None, torch_dtype=torch.float16).to("cuda")
264
+
265
+ g_cuda = None
266
+
267
+ #@markdown Can set random seed here for reproducibility.
268
+ g_cuda = torch.Generator(device='cuda')
269
+ seed = 47853 #@param {type:"number"}
270
+ g_cuda.manual_seed(seed)
271
+
272
+ #@title ##Run for generating images.
273
+
274
+ prompt = "ps1theme" #@param {type:"string"}
275
+ negative_prompt = "ugly,cartoon,3d,((disfigured)),((bad art)),(deformed)),(poorly drawn)),((blurry))" #@param {type:"string"}
276
+ num_samples = 4 #@param {type:"number"}
277
+ guidance_scale = 7.5 #@param {type:"number"}
278
+ num_inference_steps = 50 #@param {type:"number"}
279
+ height = 512 #@param {type:"number"}
280
+ width = 512 #@param {type:"number"}
281
+
282
+ with autocast("cuda"), torch.inference_mode():
283
+ images = pipe(
284
+ prompt,
285
+ height=height,
286
+ width=width,
287
+ negative_prompt=negative_prompt,
288
+ num_images_per_prompt=num_samples,
289
+ num_inference_steps=num_inference_steps,
290
+ guidance_scale=guidance_scale,
291
+ generator=g_cuda
292
+ ).images
293
+
294
+ for img in images:
295
+ display(img)
296
+
297
+ #@markdown Run Gradio UI for generating images.
298
+ import gradio as gr
299
+
300
+ def inference(prompt, negative_prompt, num_samples, height=512, width=512, num_inference_steps=50, guidance_scale=7.5):
301
+ with torch.autocast("cuda"), torch.inference_mode():
302
+ return pipe(
303
+ prompt, height=int(height), width=int(width),
304
+ negative_prompt=negative_prompt,
305
+ num_images_per_prompt=int(num_samples),
306
+ num_inference_steps=int(num_inference_steps), guidance_scale=guidance_scale,
307
+ generator=g_cuda
308
+ ).images
309
+
310
+ with gr.Blocks() as demo:
311
+ with gr.Row():
312
+ with gr.Column():
313
+ prompt = gr.Textbox(label="Prompt", value="photo of zwx dog in a bucket")
314
+ negative_prompt = gr.Textbox(label="Negative Prompt", value="")
315
+ run = gr.Button(value="Generate")
316
+ with gr.Row():
317
+ num_samples = gr.Number(label="Number of Samples", value=4)
318
+ guidance_scale = gr.Number(label="Guidance Scale", value=7.5)
319
+ with gr.Row():
320
+ height = gr.Number(label="Height", value=512)
321
+ width = gr.Number(label="Width", value=512)
322
+ num_inference_steps = gr.Slider(label="Steps", value=50)
323
+ with gr.Column():
324
+ gallery = gr.Gallery()
325
+
326
+ run.click(inference, inputs=[prompt, negative_prompt, num_samples, height, width, num_inference_steps, guidance_scale], outputs=gallery)
327
+
328
+ demo.launch(debug=True,share = True)