ginipick commited on
Commit
ca31bc0
ยท
verified ยท
1 Parent(s): 18e0d72

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +124 -89
app.py CHANGED
@@ -3,10 +3,6 @@ import re
3
  import time
4
  from os import path
5
  import tempfile
6
- import uuid
7
- import base64
8
- import mimetypes
9
- import json
10
  import io
11
  import random
12
  import string
@@ -18,7 +14,6 @@ from transformers import pipeline
18
  from safetensors.torch import load_file
19
  from huggingface_hub import hf_hub_download
20
 
21
- # Diffusers
22
  import gradio as gr
23
  from diffusers import FluxPipeline
24
 
@@ -49,6 +44,7 @@ def maybe_translate_to_english(text: str) -> str:
49
  If the prompt contains any Korean characters, translate to English.
50
  Otherwise, return as-is.
51
  """
 
52
  if re.search("[๊ฐ€-ํžฃ]", text):
53
  translated = translator(text)[0]["translation_text"]
54
  print(f"[TRANSLATE] Detected Korean -> '{text}' -> '{translated}'")
@@ -78,6 +74,7 @@ pipe = FluxPipeline.from_pretrained(
78
  torch_dtype=torch.bfloat16
79
  )
80
 
 
81
  lora_path = hf_hub_download("ByteDance/Hyper-SD", "Hyper-FLUX.1-dev-8steps-lora.safetensors")
82
  pipe.load_lora_weights(lora_path)
83
  pipe.fuse_lora(lora_scale=0.125)
@@ -93,14 +90,13 @@ def save_binary_file(file_name, data):
93
 
94
  def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
95
  """
96
- Internally modifies text within an image, returning a new image path.
97
- (Screen instructions do not mention 'Google'.)
98
  """
 
99
  api_key = os.getenv("GAPI_TOKEN", None)
100
  if not api_key:
101
- raise ValueError(
102
- "GAPI_TOKEN is missing. Please set an API key."
103
- )
104
 
105
  client = genai.Client(api_key=api_key)
106
  files = [client.files.upload(file=file_name)]
@@ -130,6 +126,7 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
130
  text_response = ""
131
  image_path = None
132
 
 
133
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
134
  temp_path = tmp.name
135
  for chunk in client.models.generate_content_stream(
@@ -137,22 +134,24 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
137
  contents=contents,
138
  config=generate_content_config,
139
  ):
140
- if not chunk.candidates or not chunk.candidates[0].content:
141
  continue
142
 
143
  candidate = chunk.candidates[0].content.parts[0]
 
144
  if candidate.inline_data:
145
  save_binary_file(temp_path, candidate.inline_data.data)
146
- print(f"[DEBUG] Returned new image -> {temp_path}")
147
  image_path = temp_path
 
148
  break
149
  else:
 
150
  text_response += chunk.text + "\n"
151
 
152
  del files
153
  return image_path, text_response
154
 
155
-
156
  #######################################
157
  # 3. Diffusion Utility
158
  #######################################
@@ -167,8 +166,7 @@ def generate_random_letters(length: int) -> str:
167
  def is_all_english(text: str) -> bool:
168
  """
169
  Check if text consists only of English letters (a-z, A-Z), digits, spaces,
170
- and a few basic punctuation characters. If so, return True.
171
- Otherwise, False (includes Korean or other characters).
172
  """
173
  import re
174
  return bool(re.match(r'^[a-zA-Z0-9\s\.,!\?\']*$', text))
@@ -176,8 +174,7 @@ def is_all_english(text: str) -> bool:
176
  def maybe_use_random_or_original(final_text: str) -> str:
177
  """
178
  If final_text is strictly English/allowed chars, use it as-is.
179
- If it contains other chars (like Korean, etc.),
180
- replace with random letters of the same length.
181
  """
182
  if not final_text:
183
  return ""
@@ -188,8 +185,7 @@ def maybe_use_random_or_original(final_text: str) -> str:
188
 
189
  def fill_prompt_with_random_texts(prompt: str, r1: str, r2: str, r3: str) -> str:
190
  """
191
- Replace <text1>, <text2>, <text3> with r1, r2, r3 respectively.
192
- <text1> is required; if missing, we append something.
193
  """
194
  if "<text1>" in prompt:
195
  prompt = prompt.replace("<text1>", r1)
@@ -200,7 +196,6 @@ def fill_prompt_with_random_texts(prompt: str, r1: str, r2: str, r3: str) -> str
200
  prompt = prompt.replace("<text2>", r2)
201
  if "<text3>" in prompt:
202
  prompt = prompt.replace("<text3>", r3)
203
-
204
  return prompt
205
 
206
  def generate_initial_image(prompt, height, width, steps, scale, seed):
@@ -219,30 +214,13 @@ def generate_initial_image(prompt, height, width, steps, scale, seed):
219
  ).images[0]
220
  return result
221
 
222
-
223
  #######################################
224
  # 4. Creating 2 Final Images
225
  #######################################
226
 
227
- def build_multi_change_instruction(r1, f1, r2, f2, r3, f3):
228
- """
229
- Summarize instructions to replace (r1->f1), (r2->f2), (r3->f3).
230
- """
231
- instructions = []
232
- if r1 and f1:
233
- instructions.append(f"Change any text reading '{r1}' in this image to '{f1}'.")
234
- if r2 and f2:
235
- instructions.append(f"Change any text reading '{r2}' in this image to '{f2}'.")
236
- if r3 and f3:
237
- instructions.append(f"Change any text reading '{r3}' in this image to '{f3}'.")
238
- if instructions:
239
- return " ".join(instructions)
240
- return "No text changes needed."
241
-
242
  def change_text_in_image_two_times(original_image, instruction):
243
  """
244
- Call the text modification function twice,
245
- returning 2 final variations.
246
  """
247
  results = []
248
  for version_tag in ["(A)", "(B)"]:
@@ -262,12 +240,13 @@ def change_text_in_image_two_times(original_image, instruction):
262
  new_img = Image.open(io.BytesIO(image_data))
263
  results.append(new_img)
264
  else:
 
 
265
  results.append(original_image)
266
  except Exception as e:
267
  raise gr.Error(f"Error: {e}")
268
  return results
269
 
270
-
271
  #######################################
272
  # 5. Main Process (Generation from Prompt)
273
  #######################################
@@ -284,51 +263,61 @@ def run_process(
284
  seed
285
  ):
286
  """
287
- 1) If prompt has Korean, translate to English
288
- 2) For each <textX>, if it's purely English, use as-is,
289
- else generate random letters of the same length.
290
- 3) Generate initial image with these placeholders
291
- 4) Then produce 2 final images by replacing placeholders with real texts
292
  """
 
293
  prompt_en = maybe_translate_to_english(prompt)
294
 
295
- # Decide random vs original for each text
296
  r1 = maybe_use_random_or_original(final_text1)
297
  r2 = maybe_use_random_or_original(final_text2)
298
  r3 = maybe_use_random_or_original(final_text3)
299
-
300
  print(f"[DEBUG] Using placeholders: r1='{r1}', r2='{r2}', r3='{r3}'")
301
 
302
- # Fill prompt
303
  final_prompt = fill_prompt_with_random_texts(prompt_en, r1, r2, r3)
304
  print(f"[DEBUG] final_prompt = {final_prompt}")
305
 
306
- # Generate initial "random/original" image
307
  _random_image = generate_initial_image(final_prompt, height, width, steps, scale, seed)
308
 
309
- # Build final instructions & call twice -> 2 final images
310
- instruction = build_multi_change_instruction(r1, final_text1, r2, final_text2, r3, final_text3)
 
 
 
 
 
 
 
 
 
311
  final_imgs = change_text_in_image_two_times(_random_image, instruction)
312
- # Return only the 2 final images
313
  return [final_imgs[0], final_imgs[1]]
314
 
315
  #######################################
316
  # 5-2. Process for Editing Uploaded Image
317
  #######################################
318
-
319
- def run_edit_process(input_image, final_text1, final_text2, final_text3):
320
  """
321
- Process for editing an uploaded image.
322
- 1) For each provided text, if it's purely English, use as-is;
323
- otherwise, replace with random letters.
324
- 2) Build instructions and call the text modification function twice.
325
  """
326
  r1 = maybe_use_random_or_original(final_text1)
327
- r2 = maybe_use_random_or_original(final_text2)
328
- r3 = maybe_use_random_or_original(final_text3)
329
- print(f"[DEBUG] Editing image with placeholders: r1='{r1}', r2='{r2}', r3='{r3}'")
330
- instruction = build_multi_change_instruction(r1, final_text1, r2, final_text2, r3, final_text3)
331
- print(f"[DEBUG] Editing instruction: {instruction}")
 
 
 
 
 
332
  final_imgs = change_text_in_image_two_times(input_image, instruction)
333
  return [final_imgs[0], final_imgs[1]]
334
 
@@ -372,16 +361,14 @@ with gr.Blocks(title="Eevery Text Imaginator: FLUX") as demo:
372
  color: #d500f9;
373
  }
374
  </style>
375
-
376
  <h2 style="text-align:center; margin-bottom: 15px;">
377
  <strong>Eevery Text Imaginator: FLUX</strong>
378
  </h2>
379
 
380
  <p style="text-align:center;">
381
  This tool generates <b>two final images</b> from a prompt
382
- containing placeholders <code>&lt;text1&gt;</code>, <code>&lt;text2&gt;</code>, <code>&lt;text3&gt;</code>.
383
- <br>If your chosen text is purely English, it will appear directly;
384
- otherwise it becomes <i>random letters</i> in the initial phase.
385
  </p>
386
 
387
  <hr style="margin: 15px 0;">
@@ -389,6 +376,9 @@ with gr.Blocks(title="Eevery Text Imaginator: FLUX") as demo:
389
  )
390
 
391
  with gr.Tabs():
 
 
 
392
  with gr.TabItem("Generate from Prompt"):
393
  with gr.Row():
394
  with gr.Column():
@@ -411,12 +401,41 @@ with gr.Blocks(title="Eevery Text Imaginator: FLUX") as demo:
411
  placeholder="(Leave blank if not used)"
412
  )
413
  with gr.Accordion("Advanced Settings (optional)", open=False):
414
- height = gr.Slider(label="Height", minimum=256, maximum=1152, step=64, value=512)
415
- width = gr.Slider(label="Width", minimum=256, maximum=1152, step=64, value=512)
416
- steps = gr.Slider(label="Inference Steps", minimum=6, maximum=25, step=1, value=8)
417
- scale = gr.Slider(label="Guidance Scale", minimum=0.0, maximum=10.0, step=0.5, value=3.5)
418
- seed = gr.Number(label="Seed", value=1234, precision=0)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
419
  run_btn = gr.Button("Generate 2 Final Images", variant="primary")
 
420
  gr.Examples(
421
  examples=[
422
  [
@@ -428,11 +447,11 @@ with gr.Blocks(title="Eevery Text Imaginator: FLUX") as demo:
428
  "ํ™˜์˜ํ•ฉ๋‹ˆ๋‹ค.", "", ""
429
  ],
430
  [
431
- "A classical poster reading <text1> in bold, as a subtitle",
432
  "้”™่ง‰", "", ""
433
  ],
434
  [
435
- "In a cartoon style, a speech bubble with <text1> and another text ",
436
  "์•ˆ๋…•", "", ""
437
  ],
438
  [
@@ -448,8 +467,16 @@ with gr.Blocks(title="Eevery Text Imaginator: FLUX") as demo:
448
  label="Example Prompts"
449
  )
450
  with gr.Column():
451
- final_image_output1 = gr.Image(label="Final Image #1", type="pil")
452
- final_image_output2 = gr.Image(label="Final Image #2", type="pil")
 
 
 
 
 
 
 
 
453
  run_btn.click(
454
  fn=run_process,
455
  inputs=[
@@ -466,33 +493,41 @@ with gr.Blocks(title="Eevery Text Imaginator: FLUX") as demo:
466
  outputs=[final_image_output1, final_image_output2]
467
  )
468
 
 
 
 
469
  with gr.TabItem("Edit Uploaded Image"):
470
  with gr.Row():
471
  with gr.Column():
 
472
  uploaded_image = gr.Image(
473
- label="Upload Image for Editing",
474
  type="pil"
475
  )
 
 
 
 
476
  final_text1_edit = gr.Textbox(
477
- label="New Text #1 (Required)",
478
  placeholder="Example: HELLO or ์•ˆ๋…•ํ•˜์„ธ์š”"
479
  )
480
- final_text2_edit = gr.Textbox(
481
- label="New Text #2 (Optional)",
482
- placeholder="Example: WORLD or ๋ฐ˜๊ฐ‘์Šต๋‹ˆ๋‹ค"
483
- )
484
- final_text3_edit = gr.Textbox(
485
- label="New Text #3 (Optional)",
486
- placeholder="(Leave blank if not used)"
487
- )
488
  run_edit_btn = gr.Button("Edit Image", variant="primary")
489
  with gr.Column():
490
- edited_image_output1 = gr.Image(label="Edited Image #1", type="pil")
491
- edited_image_output2 = gr.Image(label="Edited Image #2", type="pil")
 
 
 
 
 
 
 
 
492
  run_edit_btn.click(
493
  fn=run_edit_process,
494
- inputs=[uploaded_image, final_text1_edit, final_text2_edit, final_text3_edit],
495
  outputs=[edited_image_output1, edited_image_output2]
496
  )
497
 
498
- demo.launch(max_threads=20)
 
3
  import time
4
  from os import path
5
  import tempfile
 
 
 
 
6
  import io
7
  import random
8
  import string
 
14
  from safetensors.torch import load_file
15
  from huggingface_hub import hf_hub_download
16
 
 
17
  import gradio as gr
18
  from diffusers import FluxPipeline
19
 
 
44
  If the prompt contains any Korean characters, translate to English.
45
  Otherwise, return as-is.
46
  """
47
+ import re
48
  if re.search("[๊ฐ€-ํžฃ]", text):
49
  translated = translator(text)[0]["translation_text"]
50
  print(f"[TRANSLATE] Detected Korean -> '{text}' -> '{translated}'")
 
74
  torch_dtype=torch.bfloat16
75
  )
76
 
77
+ # ์˜ˆ์‹œ์šฉ LoRA ๋‹ค์šด๋กœ๋“œ & ํ•ฉ์น˜๊ธฐ
78
  lora_path = hf_hub_download("ByteDance/Hyper-SD", "Hyper-FLUX.1-dev-8steps-lora.safetensors")
79
  pipe.load_lora_weights(lora_path)
80
  pipe.fuse_lora(lora_scale=0.125)
 
90
 
91
  def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
92
  """
93
+ - ์ถ”๊ฐ€ ์ง€์‹œ์‚ฌํ•ญ(AIP)์„ ์ „๋‹ฌํ•ด ์ด๋ฏธ์ง€ ๊ธฐ๋ฐ˜ ํŽธ์ง‘์„ ์ˆ˜ํ–‰.
94
+ - ์‘๋‹ต์ด '์ด๋ฏธ์ง€'๋ฉด ์ €์žฅ, 'ํ…์ŠคํŠธ'๋ฉด ๋ˆ„์ ํ•˜์—ฌ ๋ฐ˜ํ™˜.
95
  """
96
+ # ๊ธฐ์กด API ํ‚ค ๋กœ์ง ์œ ์ง€ (ํ™˜๊ฒฝ ๋ณ€์ˆ˜ GAPI_TOKEN ์‚ฌ์šฉ)
97
  api_key = os.getenv("GAPI_TOKEN", None)
98
  if not api_key:
99
+ raise ValueError("GAPI_TOKEN is missing. Please set an API key.")
 
 
100
 
101
  client = genai.Client(api_key=api_key)
102
  files = [client.files.upload(file=file_name)]
 
126
  text_response = ""
127
  image_path = None
128
 
129
+ # ์ž„์‹œ ํŒŒ์ผ์— ์ด๋ฏธ์ง€ ์ €์žฅ ๊ฐ€๋Šฅํ•˜๋„๋ก ์ค€๋น„
130
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
131
  temp_path = tmp.name
132
  for chunk in client.models.generate_content_stream(
 
134
  contents=contents,
135
  config=generate_content_config,
136
  ):
137
+ if not chunk.candidates or not chunk.candidates[0].content or not chunk.candidates[0].content.parts:
138
  continue
139
 
140
  candidate = chunk.candidates[0].content.parts[0]
141
+ # ๋งŒ์•ฝ inline_data(์ด๋ฏธ์ง€ ๋ฐ์ดํ„ฐ)๊ฐ€ ์žˆ๋‹ค๋ฉด -> ์‹ค์ œ ์ด๋ฏธ์ง€ ํŽธ์ง‘ ๊ฒฐ๊ณผ
142
  if candidate.inline_data:
143
  save_binary_file(temp_path, candidate.inline_data.data)
144
+ print(f"File of mime type {candidate.inline_data.mime_type} saved to: {temp_path}")
145
  image_path = temp_path
146
+ # ์ด๋ฏธ์ง€ ํ•œ ์žฅ๋งŒ ํ™•๋ณดํ•˜๋ฉด ์ค‘๋‹จ
147
  break
148
  else:
149
+ # inline_data๊ฐ€ ์—†์œผ๋ฉด ํ…์ŠคํŠธ ๋ฐ์ดํ„ฐ์ด๋ฏ€๋กœ ๋ˆ„์ 
150
  text_response += chunk.text + "\n"
151
 
152
  del files
153
  return image_path, text_response
154
 
 
155
  #######################################
156
  # 3. Diffusion Utility
157
  #######################################
 
166
  def is_all_english(text: str) -> bool:
167
  """
168
  Check if text consists only of English letters (a-z, A-Z), digits, spaces,
169
+ and basic punctuation. If so, return True; otherwise False.
 
170
  """
171
  import re
172
  return bool(re.match(r'^[a-zA-Z0-9\s\.,!\?\']*$', text))
 
174
  def maybe_use_random_or_original(final_text: str) -> str:
175
  """
176
  If final_text is strictly English/allowed chars, use it as-is.
177
+ Else replace with random letters of the same length.
 
178
  """
179
  if not final_text:
180
  return ""
 
185
 
186
  def fill_prompt_with_random_texts(prompt: str, r1: str, r2: str, r3: str) -> str:
187
  """
188
+ Replace <text1>, <text2>, <text3> placeholders with r1, r2, r3.
 
189
  """
190
  if "<text1>" in prompt:
191
  prompt = prompt.replace("<text1>", r1)
 
196
  prompt = prompt.replace("<text2>", r2)
197
  if "<text3>" in prompt:
198
  prompt = prompt.replace("<text3>", r3)
 
199
  return prompt
200
 
201
  def generate_initial_image(prompt, height, width, steps, scale, seed):
 
214
  ).images[0]
215
  return result
216
 
 
217
  #######################################
218
  # 4. Creating 2 Final Images
219
  #######################################
220
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
221
  def change_text_in_image_two_times(original_image, instruction):
222
  """
223
+ Call the text-modification API twice, returning 2 final variations.
 
224
  """
225
  results = []
226
  for version_tag in ["(A)", "(B)"]:
 
240
  new_img = Image.open(io.BytesIO(image_data))
241
  results.append(new_img)
242
  else:
243
+ # ๋งŒ์•ฝ ์ด๋ฏธ์ง€ ์‘๋‹ต์ด ์—†๊ณ , ํ…์ŠคํŠธ๋งŒ ์˜จ ๊ฒฝ์šฐ
244
+ print("[WARNING] No image returned. text_response=", text_response)
245
  results.append(original_image)
246
  except Exception as e:
247
  raise gr.Error(f"Error: {e}")
248
  return results
249
 
 
250
  #######################################
251
  # 5. Main Process (Generation from Prompt)
252
  #######################################
 
263
  seed
264
  ):
265
  """
266
+ 1) Translate prompt if Korean -> English
267
+ 2) For each text, if not English -> random
268
+ 3) Generate initial image
269
+ 4) Replace placeholders with real text via API (2 variations)
 
270
  """
271
+ # 1) Translate prompt if needed
272
  prompt_en = maybe_translate_to_english(prompt)
273
 
274
+ # 2) Decide placeholders
275
  r1 = maybe_use_random_or_original(final_text1)
276
  r2 = maybe_use_random_or_original(final_text2)
277
  r3 = maybe_use_random_or_original(final_text3)
 
278
  print(f"[DEBUG] Using placeholders: r1='{r1}', r2='{r2}', r3='{r3}'")
279
 
280
+ # 3) Fill placeholders in prompt
281
  final_prompt = fill_prompt_with_random_texts(prompt_en, r1, r2, r3)
282
  print(f"[DEBUG] final_prompt = {final_prompt}")
283
 
284
+ # 4) Generate initial "random/original" image
285
  _random_image = generate_initial_image(final_prompt, height, width, steps, scale, seed)
286
 
287
+ # Build final instructions (replace placeholders -> real text)
288
+ instructions = []
289
+ if r1 and final_text1:
290
+ instructions.append(f"Change any text reading '{r1}' in this image to '{final_text1}'.")
291
+ if r2 and final_text2:
292
+ instructions.append(f"Change any text reading '{r2}' in this image to '{final_text2}'.")
293
+ if r3 and final_text3:
294
+ instructions.append(f"Change any text reading '{r3}' in this image to '{final_text3}'.")
295
+ instruction = " ".join(instructions) if instructions else "No text changes needed."
296
+
297
+ # Call 2 variations
298
  final_imgs = change_text_in_image_two_times(_random_image, instruction)
 
299
  return [final_imgs[0], final_imgs[1]]
300
 
301
  #######################################
302
  # 5-2. Process for Editing Uploaded Image
303
  #######################################
304
+ def run_edit_process(input_image, edit_prompt, final_text1):
 
305
  """
306
+ 1) If final_text1 is empty => skip text replacement
307
+ 2) Otherwise, combine edit_prompt + text-change instructions
308
+ 3) Call 2 times for final images
 
309
  """
310
  r1 = maybe_use_random_or_original(final_text1)
311
+ print(f"[DEBUG] Editing image with placeholder r1='{r1}'")
312
+
313
+ # *** ์ˆ˜์ • ํ•ต์‹ฌ ***
314
+ # final_text1์ด ๋น„์–ด ์žˆ์œผ๋ฉด ํ…์ŠคํŠธ ์น˜ํ™˜์„ ์ƒ๋žต,
315
+ # ์•„๋‹ˆ๋ฉด "Change any text reading 'r1' => final_text1" ๋ช…๋ น ์ถ”๊ฐ€
316
+ if not final_text1.strip():
317
+ instruction = f"{edit_prompt}"
318
+ else:
319
+ instruction = f"{edit_prompt}\nChange any text reading '{r1}' in this image to '{final_text1}'."
320
+
321
  final_imgs = change_text_in_image_two_times(input_image, instruction)
322
  return [final_imgs[0], final_imgs[1]]
323
 
 
361
  color: #d500f9;
362
  }
363
  </style>
 
364
  <h2 style="text-align:center; margin-bottom: 15px;">
365
  <strong>Eevery Text Imaginator: FLUX</strong>
366
  </h2>
367
 
368
  <p style="text-align:center;">
369
  This tool generates <b>two final images</b> from a prompt
370
+ or an uploaded image, optionally containing placeholders
371
+ <code>&lt;text1&gt;</code>, <code>&lt;text2&gt;</code>, <code>&lt;text3&gt;</code>.
 
372
  </p>
373
 
374
  <hr style="margin: 15px 0;">
 
376
  )
377
 
378
  with gr.Tabs():
379
+ ###############################################
380
+ # Tab 1) Generate from Prompt
381
+ ###############################################
382
  with gr.TabItem("Generate from Prompt"):
383
  with gr.Row():
384
  with gr.Column():
 
401
  placeholder="(Leave blank if not used)"
402
  )
403
  with gr.Accordion("Advanced Settings (optional)", open=False):
404
+ height = gr.Slider(
405
+ label="Height",
406
+ minimum=256,
407
+ maximum=1152,
408
+ step=64,
409
+ value=512
410
+ )
411
+ width = gr.Slider(
412
+ label="Width",
413
+ minimum=256,
414
+ maximum=1152,
415
+ step=64,
416
+ value=512
417
+ )
418
+ steps = gr.Slider(
419
+ label="Inference Steps",
420
+ minimum=6,
421
+ maximum=25,
422
+ step=1,
423
+ value=8
424
+ )
425
+ scale = gr.Slider(
426
+ label="Guidance Scale",
427
+ minimum=0.0,
428
+ maximum=10.0,
429
+ step=0.5,
430
+ value=3.5
431
+ )
432
+ seed = gr.Number(
433
+ label="Seed",
434
+ value=1234,
435
+ precision=0
436
+ )
437
  run_btn = gr.Button("Generate 2 Final Images", variant="primary")
438
+
439
  gr.Examples(
440
  examples=[
441
  [
 
447
  "ํ™˜์˜ํ•ฉ๋‹ˆ๋‹ค.", "", ""
448
  ],
449
  [
450
+ "A classical poster reading <text1> in bold, as a subtitle",
451
  "้”™่ง‰", "", ""
452
  ],
453
  [
454
+ "In a cartoon style, a speech bubble with <text1> and another text",
455
  "์•ˆ๋…•", "", ""
456
  ],
457
  [
 
467
  label="Example Prompts"
468
  )
469
  with gr.Column():
470
+ final_image_output1 = gr.Image(
471
+ label="Final Image #1",
472
+ type="pil"
473
+ )
474
+ final_image_output2 = gr.Image(
475
+ label="Final Image #2",
476
+ type="pil"
477
+ )
478
+
479
+ # ๋ฒ„ํŠผ ํด๋ฆญ ์‹œ ์ฒ˜๋ฆฌ
480
  run_btn.click(
481
  fn=run_process,
482
  inputs=[
 
493
  outputs=[final_image_output1, final_image_output2]
494
  )
495
 
496
+ ###############################################
497
+ # Tab 2) Edit Uploaded Image
498
+ ###############################################
499
  with gr.TabItem("Edit Uploaded Image"):
500
  with gr.Row():
501
  with gr.Column():
502
+ # Gradio ๊ตฌ๋ฒ„์ „ ํ˜ธํ™˜์„ ์œ„ํ•ด source="upload"๋Š” ์ œ๊ฑฐ
503
  uploaded_image = gr.Image(
504
+ label="Upload Image for Editing",
505
  type="pil"
506
  )
507
+ edit_prompt = gr.Textbox(
508
+ label="Additional Instruction Prompt",
509
+ placeholder="(์˜ˆ: Make the background black, add sparkles, etc.)"
510
+ )
511
  final_text1_edit = gr.Textbox(
512
+ label="Replace Text",
513
  placeholder="Example: HELLO or ์•ˆ๋…•ํ•˜์„ธ์š”"
514
  )
 
 
 
 
 
 
 
 
515
  run_edit_btn = gr.Button("Edit Image", variant="primary")
516
  with gr.Column():
517
+ edited_image_output1 = gr.Image(
518
+ label="Edited Image #1",
519
+ type="pil"
520
+ )
521
+ edited_image_output2 = gr.Image(
522
+ label="Edited Image #2",
523
+ type="pil"
524
+ )
525
+
526
+ # ์—…๋กœ๋“œ ์ด๋ฏธ์ง€ ํŽธ์ง‘ ์‹œ ์ฒ˜๋ฆฌ
527
  run_edit_btn.click(
528
  fn=run_edit_process,
529
+ inputs=[uploaded_image, edit_prompt, final_text1_edit],
530
  outputs=[edited_image_output1, edited_image_output2]
531
  )
532
 
533
+ demo.launch(max_threads=20)