Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -4,7 +4,6 @@ import random
|
|
4 |
import torch
|
5 |
from diffusers import DiffusionPipeline
|
6 |
import importlib # to import tag modules dynamically
|
7 |
-
# from tags import participant_tags, tribe_tags, role_tags, skin_tone_tags, body_type_tags, tattoo_tags, piercing_tags, expression_tags, eye_tags, hair_style_tags, position_tags, fetish_tags, location_tags, camera_tags, atmosphere_tags
|
8 |
|
9 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
10 |
model_repo_id = "John6666/wai-ani-nsfw-ponyxl-v8-sdxl" # Replace with your desired model
|
@@ -20,7 +19,6 @@ pipe.to(device)
|
|
20 |
MAX_SEED = np.iinfo(np.int32).max
|
21 |
MAX_IMAGE_SIZE = 1024
|
22 |
|
23 |
-
|
24 |
# Function to load tags dynamically based on the selected tab
|
25 |
def load_tags(active_tab):
|
26 |
if active_tab == "Gay":
|
@@ -34,7 +32,6 @@ def load_tags(active_tab):
|
|
34 |
|
35 |
return tags_module
|
36 |
|
37 |
-
|
38 |
@gradio.Interface
|
39 |
def infer(
|
40 |
prompt, negative_prompt, seed, randomize_seed, width, height, guidance_scale, num_inference_steps,
|
|
|
4 |
import torch
|
5 |
from diffusers import DiffusionPipeline
|
6 |
import importlib # to import tag modules dynamically
|
|
|
7 |
|
8 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
9 |
model_repo_id = "John6666/wai-ani-nsfw-ponyxl-v8-sdxl" # Replace with your desired model
|
|
|
19 |
MAX_SEED = np.iinfo(np.int32).max
|
20 |
MAX_IMAGE_SIZE = 1024
|
21 |
|
|
|
22 |
# Function to load tags dynamically based on the selected tab
|
23 |
def load_tags(active_tab):
|
24 |
if active_tab == "Gay":
|
|
|
32 |
|
33 |
return tags_module
|
34 |
|
|
|
35 |
@gradio.Interface
|
36 |
def infer(
|
37 |
prompt, negative_prompt, seed, randomize_seed, width, height, guidance_scale, num_inference_steps,
|