Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -38,7 +38,7 @@ with open('loras.json', 'r') as f:
|
|
38 |
|
39 |
# Initialize the base model
|
40 |
dtype = torch.bfloat16
|
41 |
-
base_model = "
|
42 |
pipe = DiffusionPipeline.from_pretrained(base_model, torch_dtype=dtype).to("cuda")
|
43 |
#pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=torch.float16).to("cuda")
|
44 |
torch.cuda.empty_cache()
|
@@ -155,7 +155,7 @@ with gr.Blocks(theme=gr.themes.Soft(), css=css) as app:
|
|
155 |
)
|
156 |
# Info blob stating what the app is running
|
157 |
info_blob = gr.HTML(
|
158 |
-
"""<div id="info_blob"> Img. Manufactory Running On:
|
159 |
)
|
160 |
|
161 |
# Info blob stating what the app is running
|
|
|
38 |
|
39 |
# Initialize the base model
|
40 |
dtype = torch.bfloat16
|
41 |
+
base_model = "shuttleai/shuttle-jaguar"
|
42 |
pipe = DiffusionPipeline.from_pretrained(base_model, torch_dtype=dtype).to("cuda")
|
43 |
#pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=torch.float16).to("cuda")
|
44 |
torch.cuda.empty_cache()
|
|
|
155 |
)
|
156 |
# Info blob stating what the app is running
|
157 |
info_blob = gr.HTML(
|
158 |
+
"""<div id="info_blob"> Img. Manufactory Running On: the Shuttle Jaguar Model (a fast open offshoot of FLUX). Nearly all of the LoRA adapters accessible via this space were trained by us in an extensive progression of inspired experiments and conceptual mini-projects. Check out our poetry translations at WWW.SILVERagePOETS.com Find our music on SoundCloud @ AlekseyCalvin & YouTube @ SilverAgePoets / AlekseyCalvin! </div>"""
|
159 |
)
|
160 |
|
161 |
# Info blob stating what the app is running
|