sunhaha123 commited on
Commit
4adaac5
·
verified ·
1 Parent(s): 519021b

Update bokeh_turboX.json

Browse files
Files changed (1) hide show
  1. bokeh_turboX.json +287 -345
bokeh_turboX.json CHANGED
@@ -1,13 +1,75 @@
1
  {
2
- "last_node_id": 65,
3
- "last_link_id": 261,
4
  "nodes": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  {
6
  "id": 30,
7
  "type": "CLIPTextEncodeSD3",
8
  "pos": {
9
- "0": 518,
10
- "1": -261,
11
  "2": 0,
12
  "3": 0,
13
  "4": 0,
@@ -22,7 +84,7 @@
22
  "1": 200
23
  },
24
  "flags": {
25
- "collapsed": true
26
  },
27
  "order": 10,
28
  "mode": 0,
@@ -63,8 +125,8 @@
63
  "Node name for S&R": "CLIPTextEncodeSD3"
64
  },
65
  "widgets_values": [
66
- "anime,render,cartoon,3d,bad structure",
67
- "anime,render,cartoon,3d,bad structure",
68
  "",
69
  "none",
70
  true,
@@ -72,12 +134,54 @@
72
  true
73
  ]
74
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75
  {
76
  "id": 32,
77
  "type": "CLIPTextEncodeSD3",
78
  "pos": {
79
- "0": 518,
80
- "1": -319,
81
  "2": 0,
82
  "3": 0,
83
  "4": 0,
@@ -92,7 +196,7 @@
92
  "1": 190
93
  },
94
  "flags": {
95
- "collapsed": true
96
  },
97
  "order": 9,
98
  "mode": 0,
@@ -141,9 +245,9 @@
141
  "Node name for S&R": "CLIPTextEncodeSD3"
142
  },
143
  "widgets_values": [
144
- "A Gothic cathedral, dimly lit environment",
145
- "A Gothic cathedral, dimly lit environment",
146
- "A Gothic cathedral, dimly lit environment",
147
  "empty_prompt",
148
  true,
149
  true,
@@ -151,11 +255,11 @@
151
  ]
152
  },
153
  {
154
- "id": 17,
155
- "type": "PrimitiveNode",
156
  "pos": {
157
- "0": 487,
158
- "1": -196,
159
  "2": 0,
160
  "3": 0,
161
  "4": 0,
@@ -166,41 +270,43 @@
166
  "9": 0
167
  },
168
  "size": {
169
- "0": 263.0819396972656,
170
- "1": 82
171
  },
172
  "flags": {},
173
- "order": 0,
174
  "mode": 0,
175
- "inputs": [],
 
 
 
 
 
 
176
  "outputs": [
177
  {
178
- "name": "INT",
179
- "type": "INT",
180
  "links": [
181
- 21
182
  ],
183
  "slot_index": 0,
184
- "widget": {
185
- "name": "seed"
186
- }
187
  }
188
  ],
189
- "title": "seed",
190
  "properties": {
191
- "Run widget replace on values": false
192
  },
193
  "widgets_values": [
194
- 443600720061490,
195
- "randomize"
196
  ]
197
  },
198
  {
199
- "id": 52,
200
- "type": "PrimitiveNode",
201
  "pos": {
202
- "0": 218,
203
- "1": -94,
204
  "2": 0,
205
  "3": 0,
206
  "4": 0,
@@ -211,43 +317,39 @@
211
  "9": 0
212
  },
213
  "size": {
214
- "0": 210,
215
- "1": 151.54025268554688
216
- },
217
- "flags": {
218
- "collapsed": false
219
  },
220
- "order": 1,
 
221
  "mode": 0,
222
  "inputs": [],
223
  "outputs": [
224
  {
225
- "name": "STRING",
226
- "type": "STRING",
227
  "links": [
228
- 222,
229
- 223
230
  ],
231
- "slot_index": 0,
232
- "widget": {
233
- "name": "clip_g"
234
- }
235
  }
236
  ],
237
- "title": "Negative_prompt",
238
  "properties": {
239
- "Run widget replace on values": false
240
  },
241
  "widgets_values": [
242
- "anime,render,cartoon,3d,bad structure"
 
 
243
  ]
244
  },
245
  {
246
- "id": 62,
247
- "type": "Note",
248
  "pos": {
249
- "0": -101,
250
- "1": 107,
251
  "2": 0,
252
  "3": 0,
253
  "4": 0,
@@ -258,27 +360,50 @@
258
  "9": 0
259
  },
260
  "size": {
261
- "0": 304.94696044921875,
262
- "1": 114.46440887451172
263
  },
264
  "flags": {},
265
- "order": 2,
266
  "mode": 0,
267
  "inputs": [],
268
- "outputs": [],
269
- "properties": {},
270
- "widgets_values": [
271
- "Do not enter overly complex prompt words, which will cause serious degradation of image performance,you can use emotional and atmospheric cues to improve the picture quality"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
272
  ],
273
- "color": "#432",
274
- "bgcolor": "#653"
 
 
 
 
275
  },
276
  {
277
- "id": 59,
278
- "type": "Note",
279
  "pos": {
280
- "0": 487,
281
- "1": 105,
282
  "2": 0,
283
  "3": 0,
284
  "4": 0,
@@ -289,27 +414,44 @@
289
  "9": 0
290
  },
291
  "size": {
292
- "0": 249.3325653076172,
293
- "1": 104.7717514038086
294
  },
295
- "flags": {},
296
- "order": 3,
 
 
297
  "mode": 0,
298
  "inputs": [],
299
- "outputs": [],
300
- "properties": {},
301
- "widgets_values": [
302
- "1920x1024 1728x1152 1152x1728 1280x1664 1440x1440"
 
 
 
 
 
 
 
 
 
 
303
  ],
304
- "color": "#432",
305
- "bgcolor": "#653"
 
 
 
 
 
306
  },
307
  {
308
- "id": 5,
309
- "type": "EmptyLatentImage",
310
  "pos": {
311
- "0": 475,
312
- "1": -54,
313
  "2": 0,
314
  "3": 0,
315
  "4": 0,
@@ -320,30 +462,35 @@
320
  "9": 0
321
  },
322
  "size": {
323
- "0": 266.8973388671875,
324
- "1": 116.21234893798828
325
  },
326
- "flags": {},
327
- "order": 4,
 
 
328
  "mode": 0,
329
  "inputs": [],
330
  "outputs": [
331
  {
332
- "name": "LATENT",
333
- "type": "LATENT",
334
  "links": [
335
- 243
 
336
  ],
337
- "slot_index": 0
 
 
 
338
  }
339
  ],
 
340
  "properties": {
341
- "Node name for S&R": "EmptyLatentImage"
342
  },
343
  "widgets_values": [
344
- 1280,
345
- 1664,
346
- 1
347
  ]
348
  },
349
  {
@@ -368,13 +515,13 @@
368
  "flags": {
369
  "collapsed": false
370
  },
371
- "order": 12,
372
  "mode": 0,
373
  "inputs": [
374
  {
375
  "name": "model",
376
  "type": "MODEL",
377
- "link": 261
378
  },
379
  {
380
  "name": "positive",
@@ -395,10 +542,6 @@
395
  "name": "seed",
396
  "type": "INT",
397
  "link": 21,
398
- "pos": [
399
- 10,
400
- 96
401
- ],
402
  "widget": {
403
  "name": "seed"
404
  }
@@ -418,121 +561,21 @@
418
  "Node name for S&R": "KSampler"
419
  },
420
  "widgets_values": [
421
- 443600720061490,
422
  "randomize",
423
- 8,
424
- 1.5,
425
  "euler_ancestral",
426
- "beta",
427
  1
428
  ]
429
  },
430
- {
431
- "id": 13,
432
- "type": "CheckpointLoaderSimple",
433
- "pos": {
434
- "0": -95,
435
- "1": -414,
436
- "2": 0,
437
- "3": 0,
438
- "4": 0,
439
- "5": 0,
440
- "6": 0,
441
- "7": 0,
442
- "8": 0,
443
- "9": 0
444
- },
445
- "size": {
446
- "0": 510.9742431640625,
447
- "1": 107.2224349975586
448
- },
449
- "flags": {},
450
- "order": 5,
451
- "mode": 0,
452
- "inputs": [],
453
- "outputs": [
454
- {
455
- "name": "MODEL",
456
- "type": "MODEL",
457
- "links": [
458
- 259
459
- ],
460
- "slot_index": 0
461
- },
462
- {
463
- "name": "CLIP",
464
- "type": "CLIP",
465
- "links": [],
466
- "slot_index": 1
467
- },
468
- {
469
- "name": "VAE",
470
- "type": "VAE",
471
- "links": [
472
- 23
473
- ],
474
- "slot_index": 2
475
- }
476
- ],
477
- "properties": {
478
- "Node name for S&R": "CheckpointLoaderSimple"
479
- },
480
- "widgets_values": [
481
- "bokeh_3.5_medium.safetensors"
482
- ]
483
- },
484
- {
485
- "id": 64,
486
- "type": "ModelSamplingSD3",
487
- "pos": {
488
- "0": 401.5684814453125,
489
- "1": -631.9461059570312,
490
- "2": 0,
491
- "3": 0,
492
- "4": 0,
493
- "5": 0,
494
- "6": 0,
495
- "7": 0,
496
- "8": 0,
497
- "9": 0
498
- },
499
- "size": {
500
- "0": 315,
501
- "1": 58
502
- },
503
- "flags": {},
504
- "order": 8,
505
- "mode": 0,
506
- "inputs": [
507
- {
508
- "name": "model",
509
- "type": "MODEL",
510
- "link": 259
511
- }
512
- ],
513
- "outputs": [
514
- {
515
- "name": "MODEL",
516
- "type": "MODEL",
517
- "links": [
518
- 260
519
- ],
520
- "slot_index": 0
521
- }
522
- ],
523
- "properties": {
524
- "Node name for S&R": "ModelSamplingSD3"
525
- },
526
- "widgets_values": [
527
- 5
528
- ]
529
- },
530
  {
531
  "id": 18,
532
  "type": "VAEDecode",
533
  "pos": {
534
- "0": 1114.0706787109375,
535
- "1": -478.5691833496094,
536
  "2": 0,
537
  "3": 0,
538
  "4": 0,
@@ -547,7 +590,7 @@
547
  "1": 50.05826187133789
548
  },
549
  "flags": {},
550
- "order": 13,
551
  "mode": 0,
552
  "inputs": [
553
  {
@@ -566,7 +609,7 @@
566
  "name": "IMAGE",
567
  "type": "IMAGE",
568
  "links": [
569
- 24
570
  ],
571
  "slot_index": 0
572
  }
@@ -576,11 +619,11 @@
576
  }
577
  },
578
  {
579
- "id": 19,
580
  "type": "PreviewImage",
581
  "pos": {
582
- "0": 1334,
583
- "1": -468,
584
  "2": 0,
585
  "3": 0,
586
  "4": 0,
@@ -590,20 +633,18 @@
590
  "8": 0,
591
  "9": 0
592
  },
593
- "size": {
594
- "0": 469.2524108886719,
595
- "1": 590.2518920898438
596
- },
597
- "flags": {
598
- "collapsed": false
599
- },
600
- "order": 14,
601
  "mode": 0,
602
  "inputs": [
603
  {
604
  "name": "images",
605
  "type": "IMAGE",
606
- "link": 24
607
  }
608
  ],
609
  "outputs": [],
@@ -612,11 +653,11 @@
612
  }
613
  },
614
  {
615
- "id": 20,
616
  "type": "PrimitiveNode",
617
  "pos": {
618
- "0": -102,
619
- "1": -94,
620
  "2": 0,
621
  "3": 0,
622
  "4": 0,
@@ -627,126 +668,35 @@
627
  "9": 0
628
  },
629
  "size": {
630
- "0": 306.11773681640625,
631
- "1": 152.6992950439453
632
  },
633
  "flags": {
634
  "collapsed": false
635
  },
636
- "order": 6,
637
  "mode": 0,
638
  "inputs": [],
639
  "outputs": [
640
  {
641
- "name": "STRING",
642
- "type": "STRING",
643
  "links": [
644
- 122,
645
- 123,
646
- 152
647
  ],
648
  "slot_index": 0,
649
  "widget": {
650
- "name": "clip_g"
651
  }
652
  }
653
  ],
654
- "title": "Positive_prompt",
655
  "properties": {
656
  "Run widget replace on values": false
657
  },
658
  "widgets_values": [
659
- "A Gothic cathedral, dimly lit environment"
660
- ]
661
- },
662
- {
663
- "id": 10,
664
- "type": "TripleCLIPLoader",
665
- "pos": {
666
- "0": -98,
667
- "1": -265,
668
- "2": 0,
669
- "3": 0,
670
- "4": 0,
671
- "5": 0,
672
- "6": 0,
673
- "7": 0,
674
- "8": 0,
675
- "9": 0
676
- },
677
- "size": {
678
- "0": 521.9664916992188,
679
- "1": 120.35124206542969
680
- },
681
- "flags": {},
682
- "order": 7,
683
- "mode": 0,
684
- "inputs": [],
685
- "outputs": [
686
- {
687
- "name": "CLIP",
688
- "type": "CLIP",
689
- "links": [
690
- 257,
691
- 258
692
- ],
693
- "slot_index": 0
694
- }
695
- ],
696
- "properties": {
697
- "Node name for S&R": "TripleCLIPLoader"
698
- },
699
- "widgets_values": [
700
- "bokeh_clip_g.safetensors",
701
- "bokeh_clip_l.safetensors",
702
- "t5xxl_fp8_e4m3fn.safetensors"
703
- ]
704
- },
705
- {
706
- "id": 65,
707
- "type": "LoraLoaderModelOnly",
708
- "pos": {
709
- "0": 805.9830932617188,
710
- "1": -651.0128784179688,
711
- "2": 0,
712
- "3": 0,
713
- "4": 0,
714
- "5": 0,
715
- "6": 0,
716
- "7": 0,
717
- "8": 0,
718
- "9": 0
719
- },
720
- "size": {
721
- "0": 315,
722
- "1": 82
723
- },
724
- "flags": {},
725
- "order": 11,
726
- "mode": 0,
727
- "inputs": [
728
- {
729
- "name": "model",
730
- "type": "MODEL",
731
- "link": 260
732
- }
733
- ],
734
- "outputs": [
735
- {
736
- "name": "MODEL",
737
- "type": "MODEL",
738
- "links": [
739
- 261
740
- ],
741
- "slot_index": 0
742
- }
743
- ],
744
- "properties": {
745
- "Node name for S&R": "LoraLoaderModelOnly"
746
- },
747
- "widgets_values": [
748
- "bokeh_8steps_turboX_lora.safetensors",
749
- 1
750
  ]
751
  }
752
  ],
@@ -775,14 +725,6 @@
775
  1,
776
  "VAE"
777
  ],
778
- [
779
- 24,
780
- 18,
781
- 0,
782
- 19,
783
- 0,
784
- "IMAGE"
785
- ],
786
  [
787
  122,
788
  20,
@@ -864,38 +806,38 @@
864
  "CLIP"
865
  ],
866
  [
867
- 259,
868
  13,
869
  0,
870
- 64,
871
  0,
872
  "MODEL"
873
  ],
874
  [
875
- 260,
876
- 64,
877
  0,
878
- 65,
879
  0,
880
  "MODEL"
881
  ],
882
  [
883
- 261,
884
- 65,
885
  0,
886
- 12,
887
  0,
888
- "MODEL"
889
  ]
890
  ],
891
  "groups": [],
892
  "config": {},
893
  "extra": {
894
  "ds": {
895
- "scale": 0.7972024500000019,
896
  "offset": [
897
- 248.55919740083328,
898
- 742.5893913836989
899
  ]
900
  }
901
  },
 
1
  {
2
+ "last_node_id": 71,
3
+ "last_link_id": 274,
4
  "nodes": [
5
+ {
6
+ "id": 62,
7
+ "type": "Note",
8
+ "pos": {
9
+ "0": -101,
10
+ "1": 107,
11
+ "2": 0,
12
+ "3": 0,
13
+ "4": 0,
14
+ "5": 0,
15
+ "6": 0,
16
+ "7": 0,
17
+ "8": 0,
18
+ "9": 0
19
+ },
20
+ "size": {
21
+ "0": 304.94696044921875,
22
+ "1": 114.46440887451172
23
+ },
24
+ "flags": {},
25
+ "order": 0,
26
+ "mode": 0,
27
+ "inputs": [],
28
+ "outputs": [],
29
+ "properties": {},
30
+ "widgets_values": [
31
+ "Do not enter overly complex prompt words, which will cause serious degradation of image performance,you can use emotional and atmospheric cues to improve the picture quality"
32
+ ],
33
+ "color": "#432",
34
+ "bgcolor": "#653"
35
+ },
36
+ {
37
+ "id": 59,
38
+ "type": "Note",
39
+ "pos": {
40
+ "0": 491,
41
+ "1": 201,
42
+ "2": 0,
43
+ "3": 0,
44
+ "4": 0,
45
+ "5": 0,
46
+ "6": 0,
47
+ "7": 0,
48
+ "8": 0,
49
+ "9": 0
50
+ },
51
+ "size": {
52
+ "0": 249.3325653076172,
53
+ "1": 104.7717514038086
54
+ },
55
+ "flags": {},
56
+ "order": 1,
57
+ "mode": 0,
58
+ "inputs": [],
59
+ "outputs": [],
60
+ "properties": {},
61
+ "widgets_values": [
62
+ "1920x1024 1728x1152 1152x1728 1280x1664 1440x1440"
63
+ ],
64
+ "color": "#432",
65
+ "bgcolor": "#653"
66
+ },
67
  {
68
  "id": 30,
69
  "type": "CLIPTextEncodeSD3",
70
  "pos": {
71
+ "0": 478,
72
+ "1": -323,
73
  "2": 0,
74
  "3": 0,
75
  "4": 0,
 
84
  "1": 200
85
  },
86
  "flags": {
87
+ "collapsed": false
88
  },
89
  "order": 10,
90
  "mode": 0,
 
125
  "Node name for S&R": "CLIPTextEncodeSD3"
126
  },
127
  "widgets_values": [
128
+ "cartoon,3d,anime,render,noise,chaos,pollution,(blurred background:0.5)",
129
+ "cartoon,3d,anime,render,noise,chaos,pollution,(blurred background:0.5)",
130
  "",
131
  "none",
132
  true,
 
134
  true
135
  ]
136
  },
137
+ {
138
+ "id": 5,
139
+ "type": "EmptyLatentImage",
140
+ "pos": {
141
+ "0": 480,
142
+ "1": 33,
143
+ "2": 0,
144
+ "3": 0,
145
+ "4": 0,
146
+ "5": 0,
147
+ "6": 0,
148
+ "7": 0,
149
+ "8": 0,
150
+ "9": 0
151
+ },
152
+ "size": {
153
+ "0": 266.8973388671875,
154
+ "1": 116.21234893798828
155
+ },
156
+ "flags": {},
157
+ "order": 2,
158
+ "mode": 0,
159
+ "inputs": [],
160
+ "outputs": [
161
+ {
162
+ "name": "LATENT",
163
+ "type": "LATENT",
164
+ "links": [
165
+ 243
166
+ ],
167
+ "slot_index": 0
168
+ }
169
+ ],
170
+ "properties": {
171
+ "Node name for S&R": "EmptyLatentImage"
172
+ },
173
+ "widgets_values": [
174
+ 1280,
175
+ 1664,
176
+ 1
177
+ ]
178
+ },
179
  {
180
  "id": 32,
181
  "type": "CLIPTextEncodeSD3",
182
  "pos": {
183
+ "0": 475,
184
+ "1": -544,
185
  "2": 0,
186
  "3": 0,
187
  "4": 0,
 
196
  "1": 190
197
  },
198
  "flags": {
199
+ "collapsed": false
200
  },
201
  "order": 9,
202
  "mode": 0,
 
245
  "Node name for S&R": "CLIPTextEncodeSD3"
246
  },
247
  "widgets_values": [
248
+ "a young japanese woman sitting on a brown couch. She is wearing a red off-shoulder sweater and black high-waisted tights. Her hair is long and black, styled with bangs, and she is wearing a white headband",
249
+ "a young japanese woman sitting on a brown couch. She is wearing a red off-shoulder sweater and black high-waisted tights. Her hair is long and black, styled with bangs, and she is wearing a white headband",
250
+ "a young japanese woman sitting on a brown couch. She is wearing a red off-shoulder sweater and black high-waisted tights. Her hair is long and black, styled with bangs, and she is wearing a white headband",
251
  "empty_prompt",
252
  true,
253
  true,
 
255
  ]
256
  },
257
  {
258
+ "id": 70,
259
+ "type": "ModelSamplingSD3",
260
  "pos": {
261
+ "0": 447,
262
+ "1": -715,
263
  "2": 0,
264
  "3": 0,
265
  "4": 0,
 
270
  "9": 0
271
  },
272
  "size": {
273
+ "0": 315,
274
+ "1": 58
275
  },
276
  "flags": {},
277
+ "order": 8,
278
  "mode": 0,
279
+ "inputs": [
280
+ {
281
+ "name": "model",
282
+ "type": "MODEL",
283
+ "link": 271
284
+ }
285
+ ],
286
  "outputs": [
287
  {
288
+ "name": "MODEL",
289
+ "type": "MODEL",
290
  "links": [
291
+ 273
292
  ],
293
  "slot_index": 0,
294
+ "shape": 3
 
 
295
  }
296
  ],
 
297
  "properties": {
298
+ "Node name for S&R": "ModelSamplingSD3"
299
  },
300
  "widgets_values": [
301
+ 5
 
302
  ]
303
  },
304
  {
305
+ "id": 10,
306
+ "type": "TripleCLIPLoader",
307
  "pos": {
308
+ "0": -98,
309
+ "1": -265,
310
  "2": 0,
311
  "3": 0,
312
  "4": 0,
 
317
  "9": 0
318
  },
319
  "size": {
320
+ "0": 521.9664916992188,
321
+ "1": 120.35124206542969
 
 
 
322
  },
323
+ "flags": {},
324
+ "order": 3,
325
  "mode": 0,
326
  "inputs": [],
327
  "outputs": [
328
  {
329
+ "name": "CLIP",
330
+ "type": "CLIP",
331
  "links": [
332
+ 257,
333
+ 258
334
  ],
335
+ "slot_index": 0
 
 
 
336
  }
337
  ],
 
338
  "properties": {
339
+ "Node name for S&R": "TripleCLIPLoader"
340
  },
341
  "widgets_values": [
342
+ "bokeh_clip_g.safetensors",
343
+ "bokeh_clip_l.safetensors",
344
+ "t5xxl_fp16.safetensors"
345
  ]
346
  },
347
  {
348
+ "id": 13,
349
+ "type": "CheckpointLoaderSimple",
350
  "pos": {
351
+ "0": -95,
352
+ "1": -414,
353
  "2": 0,
354
  "3": 0,
355
  "4": 0,
 
360
  "9": 0
361
  },
362
  "size": {
363
+ "0": 510.9742431640625,
364
+ "1": 107.2224349975586
365
  },
366
  "flags": {},
367
+ "order": 4,
368
  "mode": 0,
369
  "inputs": [],
370
+ "outputs": [
371
+ {
372
+ "name": "MODEL",
373
+ "type": "MODEL",
374
+ "links": [
375
+ 271
376
+ ],
377
+ "slot_index": 0
378
+ },
379
+ {
380
+ "name": "CLIP",
381
+ "type": "CLIP",
382
+ "links": [],
383
+ "slot_index": 1
384
+ },
385
+ {
386
+ "name": "VAE",
387
+ "type": "VAE",
388
+ "links": [
389
+ 23
390
+ ],
391
+ "slot_index": 2
392
+ }
393
  ],
394
+ "properties": {
395
+ "Node name for S&R": "CheckpointLoaderSimple"
396
+ },
397
+ "widgets_values": [
398
+ "bokeh_8steps_final.safetensors"
399
+ ]
400
  },
401
  {
402
+ "id": 20,
403
+ "type": "PrimitiveNode",
404
  "pos": {
405
+ "0": -102,
406
+ "1": -94,
407
  "2": 0,
408
  "3": 0,
409
  "4": 0,
 
414
  "9": 0
415
  },
416
  "size": {
417
+ "0": 306.11773681640625,
418
+ "1": 152.6992950439453
419
  },
420
+ "flags": {
421
+ "collapsed": false
422
+ },
423
+ "order": 5,
424
  "mode": 0,
425
  "inputs": [],
426
+ "outputs": [
427
+ {
428
+ "name": "STRING",
429
+ "type": "STRING",
430
+ "links": [
431
+ 122,
432
+ 123,
433
+ 152
434
+ ],
435
+ "slot_index": 0,
436
+ "widget": {
437
+ "name": "clip_g"
438
+ }
439
+ }
440
  ],
441
+ "title": "Positive_prompt",
442
+ "properties": {
443
+ "Run widget replace on values": false
444
+ },
445
+ "widgets_values": [
446
+ "a young japanese woman sitting on a brown couch. She is wearing a red off-shoulder sweater and black high-waisted tights. Her hair is long and black, styled with bangs, and she is wearing a white headband"
447
+ ]
448
  },
449
  {
450
+ "id": 52,
451
+ "type": "PrimitiveNode",
452
  "pos": {
453
+ "0": 218,
454
+ "1": -94,
455
  "2": 0,
456
  "3": 0,
457
  "4": 0,
 
462
  "9": 0
463
  },
464
  "size": {
465
+ "0": 210,
466
+ "1": 151.54025268554688
467
  },
468
+ "flags": {
469
+ "collapsed": false
470
+ },
471
+ "order": 6,
472
  "mode": 0,
473
  "inputs": [],
474
  "outputs": [
475
  {
476
+ "name": "STRING",
477
+ "type": "STRING",
478
  "links": [
479
+ 222,
480
+ 223
481
  ],
482
+ "slot_index": 0,
483
+ "widget": {
484
+ "name": "clip_g"
485
+ }
486
  }
487
  ],
488
+ "title": "Negative_prompt",
489
  "properties": {
490
+ "Run widget replace on values": false
491
  },
492
  "widgets_values": [
493
+ "cartoon,3d,anime,render,noise,chaos,pollution,(blurred background:0.5)"
 
 
494
  ]
495
  },
496
  {
 
515
  "flags": {
516
  "collapsed": false
517
  },
518
+ "order": 11,
519
  "mode": 0,
520
  "inputs": [
521
  {
522
  "name": "model",
523
  "type": "MODEL",
524
+ "link": 273
525
  },
526
  {
527
  "name": "positive",
 
542
  "name": "seed",
543
  "type": "INT",
544
  "link": 21,
 
 
 
 
545
  "widget": {
546
  "name": "seed"
547
  }
 
561
  "Node name for S&R": "KSampler"
562
  },
563
  "widgets_values": [
564
+ 119032793613891,
565
  "randomize",
566
+ 10,
567
+ 1.6,
568
  "euler_ancestral",
569
+ "sgm_uniform",
570
  1
571
  ]
572
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
573
  {
574
  "id": 18,
575
  "type": "VAEDecode",
576
  "pos": {
577
+ "0": 1091,
578
+ "1": -413,
579
  "2": 0,
580
  "3": 0,
581
  "4": 0,
 
590
  "1": 50.05826187133789
591
  },
592
  "flags": {},
593
+ "order": 12,
594
  "mode": 0,
595
  "inputs": [
596
  {
 
609
  "name": "IMAGE",
610
  "type": "IMAGE",
611
  "links": [
612
+ 274
613
  ],
614
  "slot_index": 0
615
  }
 
619
  }
620
  },
621
  {
622
+ "id": 71,
623
  "type": "PreviewImage",
624
  "pos": {
625
+ "0": 1332,
626
+ "1": -504,
627
  "2": 0,
628
  "3": 0,
629
  "4": 0,
 
633
  "8": 0,
634
  "9": 0
635
  },
636
+ "size": [
637
+ 373.8262602387847,
638
+ 537.5194426782502
639
+ ],
640
+ "flags": {},
641
+ "order": 13,
 
 
642
  "mode": 0,
643
  "inputs": [
644
  {
645
  "name": "images",
646
  "type": "IMAGE",
647
+ "link": 274
648
  }
649
  ],
650
  "outputs": [],
 
653
  }
654
  },
655
  {
656
+ "id": 17,
657
  "type": "PrimitiveNode",
658
  "pos": {
659
+ "0": 504,
660
+ "1": -74,
661
  "2": 0,
662
  "3": 0,
663
  "4": 0,
 
668
  "9": 0
669
  },
670
  "size": {
671
+ "0": 263.0819396972656,
672
+ "1": 82
673
  },
674
  "flags": {
675
  "collapsed": false
676
  },
677
+ "order": 7,
678
  "mode": 0,
679
  "inputs": [],
680
  "outputs": [
681
  {
682
+ "name": "INT",
683
+ "type": "INT",
684
  "links": [
685
+ 21
 
 
686
  ],
687
  "slot_index": 0,
688
  "widget": {
689
+ "name": "seed"
690
  }
691
  }
692
  ],
693
+ "title": "seed",
694
  "properties": {
695
  "Run widget replace on values": false
696
  },
697
  "widgets_values": [
698
+ 119032793613891,
699
+ "increment"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
700
  ]
701
  }
702
  ],
 
725
  1,
726
  "VAE"
727
  ],
 
 
 
 
 
 
 
 
728
  [
729
  122,
730
  20,
 
806
  "CLIP"
807
  ],
808
  [
809
+ 271,
810
  13,
811
  0,
812
+ 70,
813
  0,
814
  "MODEL"
815
  ],
816
  [
817
+ 273,
818
+ 70,
819
  0,
820
+ 12,
821
  0,
822
  "MODEL"
823
  ],
824
  [
825
+ 274,
826
+ 18,
827
  0,
828
+ 71,
829
  0,
830
+ "IMAGE"
831
  ]
832
  ],
833
  "groups": [],
834
  "config": {},
835
  "extra": {
836
  "ds": {
837
+ "scale": 1.0610764609500207,
838
  "offset": [
839
+ -17.273219088388544,
840
+ 624.0579759871756
841
  ]
842
  }
843
  },