samanthavaca commited on
Commit
0083054
·
verified ·
1 Parent(s): ce0586b

End of training

Browse files
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ library_name: diffusers
4
+ tags:
5
+ - stable-diffusion
6
+ - stable-diffusion-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - controlnet
10
+ - diffusers-training
11
+ base_model: runwayml/stable-diffusion-v1-5
12
+ inference: true
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+
19
+ # controlnet-samanthavaca/qrcodemodeljulyeleven
20
+
21
+ These are controlnet weights trained on runwayml/stable-diffusion-v1-5 with new type of conditioning.
22
+ You can find some example images below.
23
+
24
+ prompt: A beautiful magical forest during early dawn in the style of Bambi.
25
+ ![images_0)](./images_0.png)
26
+ prompt: A kingdom in the middle of a hot desert in the style of Aladdin.
27
+ ![images_1)](./images_1.png)
28
+ prompt: A huge city like New York City at night in the style of Big Hero 6.
29
+ ![images_2)](./images_2.png)
30
+
31
+
32
+
33
+ ## Intended uses & limitations
34
+
35
+ #### How to use
36
+
37
+ ```python
38
+ # TODO: add an example code snippet for running this diffusion pipeline
39
+ ```
40
+
41
+ #### Limitations and bias
42
+
43
+ [TODO: provide examples of latent issues and potential remediations]
44
+
45
+ ## Training details
46
+
47
+ [TODO: describe the data used to train the model]
config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "ControlNetModel",
3
+ "_diffusers_version": "0.30.0.dev0",
4
+ "_name_or_path": "monster-labs/control_v1p_sd15_qrcode_monster",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": null,
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": null,
9
+ "attention_head_dim": 8,
10
+ "block_out_channels": [
11
+ 320,
12
+ 640,
13
+ 1280,
14
+ 1280
15
+ ],
16
+ "class_embed_type": null,
17
+ "conditioning_channels": 3,
18
+ "conditioning_embedding_out_channels": [
19
+ 16,
20
+ 32,
21
+ 96,
22
+ 256
23
+ ],
24
+ "controlnet_conditioning_channel_order": "rgb",
25
+ "cross_attention_dim": 768,
26
+ "down_block_types": [
27
+ "CrossAttnDownBlock2D",
28
+ "CrossAttnDownBlock2D",
29
+ "CrossAttnDownBlock2D",
30
+ "DownBlock2D"
31
+ ],
32
+ "downsample_padding": 1,
33
+ "encoder_hid_dim": null,
34
+ "encoder_hid_dim_type": null,
35
+ "flip_sin_to_cos": true,
36
+ "freq_shift": 0,
37
+ "global_pool_conditions": false,
38
+ "in_channels": 4,
39
+ "layers_per_block": 2,
40
+ "mid_block_scale_factor": 1,
41
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
42
+ "norm_eps": 1e-05,
43
+ "norm_num_groups": 32,
44
+ "num_attention_heads": null,
45
+ "num_class_embeds": null,
46
+ "only_cross_attention": false,
47
+ "projection_class_embeddings_input_dim": null,
48
+ "resnet_time_scale_shift": "default",
49
+ "transformer_layers_per_block": 1,
50
+ "upcast_attention": false,
51
+ "use_linear_projection": false
52
+ }
diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f08ba960bfaf6a2589059c9bd189f061e483f4a30e9890a9d7c756d269f4a4ec
3
+ size 1445157120
image_control.png ADDED
images_0.png ADDED
images_1.png ADDED
images_2.png ADDED