Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -28,6 +28,22 @@ current_dir = os.path.dirname(os.path.abspath(__file__))
|
|
28 |
comfyui_path = os.path.join(current_dir, "ComfyUI")
|
29 |
sys.path.append(comfyui_path)
|
30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
from nodes import NODE_CLASS_MAPPINGS
|
32 |
from comfy import model_management
|
33 |
import folder_paths
|
@@ -47,6 +63,7 @@ def get_value_at_index(obj: Union[Sequence, Mapping], index: int) -> Any:
|
|
47 |
|
48 |
# Baixar modelos necessários
|
49 |
def download_models():
|
|
|
50 |
models = [
|
51 |
("black-forest-labs/FLUX.1-Redux-dev", "flux1-redux-dev.safetensors", "models/style_models"),
|
52 |
("comfyanonymous/flux_text_encoders", "t5xxl_fp16.safetensors", "models/text_encoders"),
|
@@ -60,6 +77,9 @@ def download_models():
|
|
60 |
for repo_id, filename, local_dir in models:
|
61 |
hf_hub_download(repo_id=repo_id, filename=filename, local_dir=local_dir)
|
62 |
|
|
|
|
|
|
|
63 |
# Inicializar modelos
|
64 |
print("Inicializando modelos...")
|
65 |
with torch.inference_mode():
|
@@ -87,22 +107,9 @@ with torch.inference_mode():
|
|
87 |
]
|
88 |
model_management.load_models_gpu(valid_models)
|
89 |
|
90 |
-
def import_custom_nodes():
|
91 |
-
import asyncio
|
92 |
-
import execution
|
93 |
-
from nodes import init_extra_nodes
|
94 |
-
import server
|
95 |
-
|
96 |
-
loop = asyncio.new_event_loop()
|
97 |
-
asyncio.set_event_loop(loop)
|
98 |
-
server_instance = server.PromptServer(loop)
|
99 |
-
execution.PromptQueue(server_instance)
|
100 |
-
init_extra_nodes()
|
101 |
-
|
102 |
@spaces.GPU
|
103 |
def generate_image(prompt, input_image, lora_weight, progress=gr.Progress(track_tqdm=True)):
|
104 |
"""Função principal de geração com monitoramento de progresso"""
|
105 |
-
import_custom_nodes()
|
106 |
try:
|
107 |
with torch.inference_mode():
|
108 |
# Codificar texto
|
@@ -161,7 +168,4 @@ with gr.Blocks() as app:
|
|
161 |
)
|
162 |
|
163 |
if __name__ == "__main__":
|
164 |
-
# Download models at startup
|
165 |
-
download_models()
|
166 |
-
# Launch the app
|
167 |
app.launch()
|
|
|
28 |
comfyui_path = os.path.join(current_dir, "ComfyUI")
|
29 |
sys.path.append(comfyui_path)
|
30 |
|
31 |
+
# Import custom nodes primeiro
|
32 |
+
def import_custom_nodes():
|
33 |
+
import asyncio
|
34 |
+
import execution
|
35 |
+
from nodes import init_extra_nodes
|
36 |
+
import server
|
37 |
+
|
38 |
+
loop = asyncio.new_event_loop()
|
39 |
+
asyncio.set_event_loop(loop)
|
40 |
+
server_instance = server.PromptServer(loop)
|
41 |
+
execution.PromptQueue(server_instance)
|
42 |
+
init_extra_nodes()
|
43 |
+
|
44 |
+
print("Importando nós customizados...")
|
45 |
+
import_custom_nodes()
|
46 |
+
|
47 |
from nodes import NODE_CLASS_MAPPINGS
|
48 |
from comfy import model_management
|
49 |
import folder_paths
|
|
|
63 |
|
64 |
# Baixar modelos necessários
|
65 |
def download_models():
|
66 |
+
print("Baixando modelos...")
|
67 |
models = [
|
68 |
("black-forest-labs/FLUX.1-Redux-dev", "flux1-redux-dev.safetensors", "models/style_models"),
|
69 |
("comfyanonymous/flux_text_encoders", "t5xxl_fp16.safetensors", "models/text_encoders"),
|
|
|
77 |
for repo_id, filename, local_dir in models:
|
78 |
hf_hub_download(repo_id=repo_id, filename=filename, local_dir=local_dir)
|
79 |
|
80 |
+
# Download models antes de inicializar
|
81 |
+
download_models()
|
82 |
+
|
83 |
# Inicializar modelos
|
84 |
print("Inicializando modelos...")
|
85 |
with torch.inference_mode():
|
|
|
107 |
]
|
108 |
model_management.load_models_gpu(valid_models)
|
109 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
110 |
@spaces.GPU
|
111 |
def generate_image(prompt, input_image, lora_weight, progress=gr.Progress(track_tqdm=True)):
|
112 |
"""Função principal de geração com monitoramento de progresso"""
|
|
|
113 |
try:
|
114 |
with torch.inference_mode():
|
115 |
# Codificar texto
|
|
|
168 |
)
|
169 |
|
170 |
if __name__ == "__main__":
|
|
|
|
|
|
|
171 |
app.launch()
|