Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -83,6 +83,12 @@ with torch.inference_mode():
|
|
83 |
type="flux"
|
84 |
)
|
85 |
|
|
|
|
|
|
|
|
|
|
|
|
|
86 |
# Style Model
|
87 |
stylemodelloader = NODE_CLASS_MAPPINGS["StyleModelLoader"]()
|
88 |
stylemodelloader_441 = stylemodelloader.load_style_model(
|
@@ -127,6 +133,7 @@ def generate_image(prompt, input_image, lora_weight, guidance, downsampling_fact
|
|
127 |
weight=weight,
|
128 |
conditioning=flux_guidance[0],
|
129 |
style_model=stylemodelloader_441[0],
|
|
|
130 |
image=loaded_image[0]
|
131 |
)
|
132 |
|
|
|
83 |
type="flux"
|
84 |
)
|
85 |
|
86 |
+
# CLIP Vision
|
87 |
+
clipvisionloader = NODE_CLASS_MAPPINGS["CLIPVisionLoader"]()
|
88 |
+
clip_vision = clipvisionloader.load_clip(
|
89 |
+
clip_name="model.safetensors"
|
90 |
+
)
|
91 |
+
|
92 |
# Style Model
|
93 |
stylemodelloader = NODE_CLASS_MAPPINGS["StyleModelLoader"]()
|
94 |
stylemodelloader_441 = stylemodelloader.load_style_model(
|
|
|
133 |
weight=weight,
|
134 |
conditioning=flux_guidance[0],
|
135 |
style_model=stylemodelloader_441[0],
|
136 |
+
clip_vision=clip_vision[0],
|
137 |
image=loaded_image[0]
|
138 |
)
|
139 |
|