Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -4,7 +4,6 @@ import random
|
|
| 4 |
import torch
|
| 5 |
from diffusers import DiffusionPipeline
|
| 6 |
import importlib # to import tag modules dynamically
|
| 7 |
-
# from tags import participant_tags, tribe_tags, role_tags, skin_tone_tags, body_type_tags, tattoo_tags, piercing_tags, expression_tags, eye_tags, hair_style_tags, position_tags, fetish_tags, location_tags, camera_tags, atmosphere_tags
|
| 8 |
|
| 9 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 10 |
model_repo_id = "John6666/wai-ani-nsfw-ponyxl-v8-sdxl" # Replace with your desired model
|
|
@@ -20,7 +19,6 @@ pipe.to(device)
|
|
| 20 |
MAX_SEED = np.iinfo(np.int32).max
|
| 21 |
MAX_IMAGE_SIZE = 1024
|
| 22 |
|
| 23 |
-
|
| 24 |
# Function to load tags dynamically based on the selected tab
|
| 25 |
def load_tags(active_tab):
|
| 26 |
if active_tab == "Gay":
|
|
@@ -34,7 +32,6 @@ def load_tags(active_tab):
|
|
| 34 |
|
| 35 |
return tags_module
|
| 36 |
|
| 37 |
-
|
| 38 |
@gradio.Interface
|
| 39 |
def infer(
|
| 40 |
prompt, negative_prompt, seed, randomize_seed, width, height, guidance_scale, num_inference_steps,
|
|
|
|
| 4 |
import torch
|
| 5 |
from diffusers import DiffusionPipeline
|
| 6 |
import importlib # to import tag modules dynamically
|
|
|
|
| 7 |
|
| 8 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 9 |
model_repo_id = "John6666/wai-ani-nsfw-ponyxl-v8-sdxl" # Replace with your desired model
|
|
|
|
| 19 |
MAX_SEED = np.iinfo(np.int32).max
|
| 20 |
MAX_IMAGE_SIZE = 1024
|
| 21 |
|
|
|
|
| 22 |
# Function to load tags dynamically based on the selected tab
|
| 23 |
def load_tags(active_tab):
|
| 24 |
if active_tab == "Gay":
|
|
|
|
| 32 |
|
| 33 |
return tags_module
|
| 34 |
|
|
|
|
| 35 |
@gradio.Interface
|
| 36 |
def infer(
|
| 37 |
prompt, negative_prompt, seed, randomize_seed, width, height, guidance_scale, num_inference_steps,
|