ginipick commited on
Commit
0399de8
ยท
verified ยท
1 Parent(s): 819fc44

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +173 -91
app.py CHANGED
@@ -35,16 +35,15 @@ os.environ["TRANSFORMERS_CACHE"] = CACHE_PATH
35
  os.environ["HF_HUB_CACHE"] = CACHE_PATH
36
  os.environ["HF_HOME"] = CACHE_PATH
37
 
38
- # ํƒ€์ด๋จธ ํด๋ž˜์Šค
39
  class timer:
40
  def __init__(self, method_name="timed process"):
41
  self.method = method_name
42
  def __enter__(self):
43
  self.start = time.time()
44
- print(f"{self.method} starts")
45
  def __exit__(self, exc_type, exc_val, exc_tb):
46
  end = time.time()
47
- print(f"{self.method} took {str(round(end - self.start, 2))}s")
48
 
49
  #######################################
50
  # 1. FLUX ํŒŒ์ดํ”„๋ผ์ธ ๋กœ๋“œ
@@ -61,30 +60,23 @@ pipe = FluxPipeline.from_pretrained(
61
  lora_path = hf_hub_download("ByteDance/Hyper-SD", "Hyper-FLUX.1-dev-8steps-lora.safetensors")
62
  pipe.load_lora_weights(lora_path)
63
  pipe.fuse_lora(lora_scale=0.125)
64
-
65
  pipe.to(device="cuda", dtype=torch.bfloat16)
66
 
67
  #######################################
68
- # 2. Google GenAI ๋ชจ๋ธ๋กœ ํ…์ŠคํŠธ ๋ณ€ํ™˜ ํ•จ์ˆ˜
69
  #######################################
70
 
71
  def save_binary_file(file_name, data):
72
- """Google GenAI์—์„œ ์‘๋‹ต๋ฐ›์€ ์ด์ง„ ๋ฐ์ดํ„ฐ๋ฅผ ์ด๋ฏธ์ง€ ํŒŒ์ผ๋กœ ์ €์žฅ"""
73
  with open(file_name, "wb") as f:
74
  f.write(data)
75
 
76
  def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
77
- """
78
- Google GenAI(gemini) ๋ชจ๋ธ์„ ํ†ตํ•ด ์ด๋ฏธ์ง€/ํ…์ŠคํŠธ๋ฅผ ์ƒ์„ฑํ•˜๊ฑฐ๋‚˜ ๋ณ€ํ™˜.
79
- - text: ๋ณ€๊ฒฝํ•  ํ…์ŠคํŠธ๋‚˜ ๋ช…๋ น์–ด ๋“ฑ ํ”„๋กฌํ”„ํŠธ
80
- - file_name: ์›๋ณธ ์ด๋ฏธ์ง€(์˜ˆ: .png) ๊ฒฝ๋กœ
81
- - model: ์‚ฌ์šฉํ•  gemini ๋ชจ๋ธ ์ด๋ฆ„
82
- """
83
  api_key = os.getenv("GAPI_TOKEN", None)
84
  if not api_key:
85
  raise ValueError(
86
  "GAPI_TOKEN ํ™˜๊ฒฝ ๋ณ€์ˆ˜๊ฐ€ ์„ค์ •๋˜์ง€ ์•Š์•˜์Šต๋‹ˆ๋‹ค. "
87
- "Google GenAI API ์‚ฌ์šฉ์„ ์œ„ํ•ด์„œ๋Š” GAPI_TOKEN์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค."
88
  )
89
 
90
  client = genai.Client(api_key=api_key)
@@ -122,13 +114,13 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
122
  contents=contents,
123
  config=generate_content_config,
124
  ):
125
- if not chunk.candidates or not chunk.candidates[0].content or not chunk.candidates[0].content.parts:
126
  continue
127
- candidate = chunk.candidates[0].content.parts[0]
128
 
 
129
  if candidate.inline_data:
130
  save_binary_file(temp_path, candidate.inline_data.data)
131
- print(f"File of mime type {candidate.inline_data.mime_type} saved to: {temp_path}")
132
  image_path = temp_path
133
  break
134
  else:
@@ -138,24 +130,43 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
138
  return image_path, text_response
139
 
140
  #######################################
141
- # 3. Diffusion + GoogleGenAI๋ฅผ ์—ฐ๊ฒฐ
142
  #######################################
143
 
144
- def generate_initial_image(prompt, text, height, width, steps, scale, seed):
 
 
 
 
 
145
  """
146
- 1) FLUX ํŒŒ์ดํ”„๋ผ์ธ์„ ์‚ฌ์šฉํ•ด 'text'๊ฐ€ ๋“ค์–ด๊ฐ„ ์ด๋ฏธ์ง€๋ฅผ ์ƒ์„ฑ
147
- - prompt ๋‚ด <text>๊ฐ€ ์žˆ์œผ๋ฉด ์น˜ํ™˜, ์—†์œผ๋ฉด ์ž๋™ ์ถ”๊ฐ€
 
 
148
  """
149
- if "<text>" in prompt:
150
- combined_prompt = prompt.replace("<text>", text)
 
151
  else:
152
- combined_prompt = f"{prompt} with clear readable text that says '{text}'"
153
-
154
- print("[DEBUG] combined_prompt:", combined_prompt)
 
 
 
 
 
 
 
155
 
156
- with torch.inference_mode(), torch.autocast("cuda", dtype=torch.bfloat16), timer("GenerateInitialImage"):
 
 
 
 
157
  result = pipe(
158
- prompt=[combined_prompt],
159
  generator=torch.Generator().manual_seed(int(seed)),
160
  num_inference_steps=int(steps),
161
  guidance_scale=float(scale),
@@ -163,108 +174,163 @@ def generate_initial_image(prompt, text, height, width, steps, scale, seed):
163
  width=int(width),
164
  max_sequence_length=256
165
  ).images[0]
166
-
167
  return result
168
 
169
- def change_text_in_image(original_image, new_text):
 
170
  """
171
- 2) Gemini ๋ชจ๋ธ์„ ํ†ตํ•ด,
172
- original_image ๋‚ด ํ…์ŠคํŠธ๋ฅผ `new_text`๋กœ ๋ณ€๊ฒฝํ•œ ์ด๋ฏธ์ง€ ๋ฐ˜ํ™˜
173
  """
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
174
  try:
 
175
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
176
  original_path = tmp.name
177
  original_image.save(original_path)
178
 
179
  image_path, text_response = generate_by_google_genai(
180
- text=f"Change the text in this image to: '{new_text}'",
181
  file_name=original_path
182
  )
183
-
184
  if image_path:
185
  with open(image_path, "rb") as f:
186
  image_data = f.read()
187
- modified_img = Image.open(io.BytesIO(image_data))
188
- return modified_img
189
  else:
190
- # ์ด๋ฏธ์ง€๊ฐ€ ์—†์ด ํ…์ŠคํŠธ๋งŒ ์‘๋‹ต๋œ ๊ฒฝ์šฐ
191
- return None
192
-
193
  except Exception as e:
194
  raise gr.Error(f"Error: {e}")
195
 
196
  #######################################
197
- # 4. ์ž„์˜ ์•ŒํŒŒ๋ฒณ ์ƒ์„ฑ
198
- #######################################
199
-
200
- def generate_random_letters(length: int) -> str:
201
- """
202
- length ๊ธธ์ด์˜ ์ž„์˜ ์•ŒํŒŒ๋ฒณ(๋Œ€์†Œ๋ฌธ์ž) ๋ฌธ์ž์—ด ์ƒ์„ฑ
203
- """
204
- letters = string.ascii_lowercase + string.ascii_uppercase
205
- return "".join(random.choice(letters) for _ in range(length))
206
-
207
- #######################################
208
- # 5. ์ตœ์ข… ํ•จ์ˆ˜: ๋ฒ„ํŠผ ํ•œ ๋ฒˆ์œผ๋กœ
209
- # (1) ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ์œผ๋กœ 1์ฐจ ์ด๋ฏธ์ง€ ์ƒ์„ฑ
210
- # (2) ์ง„์งœ "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"๋กœ 2์ฐจ ์ด๋ฏธ์ง€ ์ƒ์„ฑ
211
  #######################################
212
 
213
- def run_full_process(prompt, final_text, height, width, steps, scale, seed):
 
 
 
 
 
 
 
 
 
 
214
  """
215
- - final_text์˜ ๊ธธ์ด์— ๋งž์ถฐ ๋žœ๋ค ์•ŒํŒŒ๋ฒณ์„ ์ƒ์„ฑ -> 1์ฐจ ์ด๋ฏธ์ง€
216
- - ๊ทธ 1์ฐจ ์ด๋ฏธ์ง€๋ฅผ ๋ฐ”ํƒ•์œผ๋กœ, final_text๋กœ ๊ต์ฒด -> 2์ฐจ ์ตœ์ข… ์ด๋ฏธ์ง€
 
217
  """
218
- # (A) ์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ(final_text) ๊ธ€์ž์ˆ˜๋งŒํผ ์ž„์˜ ์•ŒํŒŒ๋ฒณ ์ƒ์„ฑ
219
- random_len = len(final_text)
220
- random_text = generate_random_letters(random_len)
221
- print(f"[STEP] final_text='{final_text}' => random_text='{random_text}'")
222
-
223
- # (B) 1์ฐจ ์ด๋ฏธ์ง€: ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ์œผ๋กœ ์ƒ์„ฑ
224
- random_image = generate_initial_image(prompt, random_text, height, width, steps, scale, seed)
 
 
 
 
 
 
 
 
 
 
 
 
225
 
226
- # (C) 2์ฐจ ์ด๋ฏธ์ง€: ์‹ค์ œ final_text๋กœ ๊ต์ฒด
227
- final_image = change_text_in_image(random_image, final_text)
228
-
229
  return [random_image, final_image]
230
 
231
  #######################################
232
- # 6. Gradio UI
233
  #######################################
234
 
235
- with gr.Blocks(title="Flux + Google GenAI (Random & Then Real Text)") as demo:
236
  gr.Markdown(
237
  """
238
- # Flux + Google GenAI: ๋‘ ๋‹จ๊ณ„์— ๊ฑธ์นœ ํ…์ŠคํŠธ ๊ต์ฒด
239
-
240
- **์‚ฌ์šฉ ํ๋ฆ„**
241
- 1) Prompt์— ์žฅ๋ฉด์ด๋‚˜ ์Šคํƒ€์ผ์„ ์ž‘์„ฑ (ํ•„์š”ํ•˜๋ฉด `<text>` ๊ตฌ๋ถ„์ž ์‚ฌ์šฉ)
242
- 2) "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ" ์— ์ตœ์ข… ์›ํ•˜๋Š” ๋ฌธ์ž์—ด์„ ์ž…๋ ฅ (์˜ˆ: "์•ˆ๋…•ํ•˜์„ธ์š”")
243
- 3) "Generate Images" ๋ฒ„ํŠผ์„ ๋ˆ„๋ฅด๋ฉด,
244
- - ๋จผ์ € "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ" ๊ธธ์ด์— ๋งž๋Š” **๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ**์„ ๋„ฃ์–ด ์ด๋ฏธ์ง€ ์ƒ์„ฑ (1์ฐจ ์ด๋ฏธ์ง€)
245
- - ์ด์–ด์„œ **์ง„์งœ** "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"๋กœ ๋‹ค์‹œ ๊ต์ฒด(2์ฐจ ์ตœ์ข… ์ด๋ฏธ์ง€)
246
- 4) ๊ฒฐ๊ณผ๋กœ ๋‘ ์žฅ์˜ ์ด๋ฏธ์ง€๋ฅผ ํ™•์ธํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
247
-
248
- **์ฃผ์š” ํฌ์ธํŠธ**
249
- - "์ด๋ฏธ์ง€ ์•ˆ์— ๋“ค์–ด๊ฐˆ ํ…์ŠคํŠธ"๋Š” **UI์— ๋…ธ์ถœ๋˜์ง€ ์•Š์œผ๋ฉฐ**(์‚ฌ์šฉ์ž ์ž…๋ ฅ ๋ถˆ๊ฐ€), ์˜ค์ง ๋‚ด๋ถ€์—์„œ ์ž๋™ ์„ค์ •๋ฉ๋‹ˆ๋‹ค.
250
- - 1์ฐจ ์ด๋ฏธ์ง€๋Š” ์™„์ „ํžˆ ์ž„์˜์˜ ์•ŒํŒŒ๋ฒณ ํ…์ŠคํŠธ๋ฅผ ํฌํ•จํ•ฉ๋‹ˆ๋‹ค.
251
- - 2์ฐจ ์ด๋ฏธ์ง€๋Š” ์ตœ์ข…์ ์œผ๋กœ ์‚ฌ์šฉ์ž๊ฐ€ ์ž…๋ ฅํ•œ "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"๋ฅผ ํฌํ•จํ•ฉ๋‹ˆ๋‹ค.
 
252
  """
253
  )
254
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
255
  with gr.Row():
256
  with gr.Column():
257
  prompt_input = gr.Textbox(
258
  lines=3,
259
- label="Prompt (use `<text>` if you want)",
260
- placeholder="e.g. A white cat with a speech bubble <text>"
 
 
 
 
261
  )
262
- final_text_input = gr.Textbox(
263
- lines=1,
264
- label="์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ",
265
- placeholder="์˜ˆ) ์•ˆ๋…•ํ•˜์„ธ์š”"
266
  )
267
- with gr.Accordion("๊ณ ๊ธ‰ ์„ค์ • (ํ™•์žฅ)", open=False):
 
 
 
 
 
268
  height = gr.Slider(label="Height", minimum=256, maximum=1152, step=64, value=512)
269
  width = gr.Slider(label="Width", minimum=256, maximum=1152, step=64, value=512)
270
  steps = gr.Slider(label="Inference Steps", minimum=6, maximum=25, step=1, value=8)
@@ -273,14 +339,30 @@ with gr.Blocks(title="Flux + Google GenAI (Random & Then Real Text)") as demo:
273
 
274
  run_btn = gr.Button("Generate Images", variant="primary")
275
 
 
 
 
 
 
 
276
  with gr.Column():
277
  random_image_output = gr.Image(label="1) Random Text Image", type="pil")
278
  final_image_output = gr.Image(label="2) Final Text Image", type="pil")
279
 
280
- # ๋ฒ„ํŠผ ์•ก์…˜: ์œ„ ๋‹จ๊ณ„๋“ค์„ ๋ชจ๋‘ ์‹คํ–‰ -> ๊ฒฐ๊ณผ 2์žฅ ์ถœ๋ ฅ
281
  run_btn.click(
282
- fn=run_full_process,
283
- inputs=[prompt_input, final_text_input, height, width, steps, scale, seed],
 
 
 
 
 
 
 
 
 
 
284
  outputs=[random_image_output, final_image_output]
285
  )
286
 
 
35
  os.environ["HF_HUB_CACHE"] = CACHE_PATH
36
  os.environ["HF_HOME"] = CACHE_PATH
37
 
 
38
  class timer:
39
  def __init__(self, method_name="timed process"):
40
  self.method = method_name
41
  def __enter__(self):
42
  self.start = time.time()
43
+ print(f"[TIMER] {self.method} starts")
44
  def __exit__(self, exc_type, exc_val, exc_tb):
45
  end = time.time()
46
+ print(f"[TIMER] {self.method} took {round(end - self.start, 2)}s")
47
 
48
  #######################################
49
  # 1. FLUX ํŒŒ์ดํ”„๋ผ์ธ ๋กœ๋“œ
 
60
  lora_path = hf_hub_download("ByteDance/Hyper-SD", "Hyper-FLUX.1-dev-8steps-lora.safetensors")
61
  pipe.load_lora_weights(lora_path)
62
  pipe.fuse_lora(lora_scale=0.125)
 
63
  pipe.to(device="cuda", dtype=torch.bfloat16)
64
 
65
  #######################################
66
+ # 2. Google GenAI (Gemini) - ์ด๋ฏธ์ง€ ๋ณ€ํ™˜ ํ•จ์ˆ˜
67
  #######################################
68
 
69
  def save_binary_file(file_name, data):
 
70
  with open(file_name, "wb") as f:
71
  f.write(data)
72
 
73
  def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
74
+ """Gemini ๋ชจ๋ธ์„ ํ†ตํ•ด ์ด๋ฏธ์ง€ ๋‚ด๋ถ€ ํ…์ŠคํŠธ๋ฅผ ๋ณ€๊ฒฝ."""
 
 
 
 
 
75
  api_key = os.getenv("GAPI_TOKEN", None)
76
  if not api_key:
77
  raise ValueError(
78
  "GAPI_TOKEN ํ™˜๊ฒฝ ๋ณ€์ˆ˜๊ฐ€ ์„ค์ •๋˜์ง€ ์•Š์•˜์Šต๋‹ˆ๋‹ค. "
79
+ "Google GenAI API๋ฅผ ์‚ฌ์šฉํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” GAPI_TOKEN์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค."
80
  )
81
 
82
  client = genai.Client(api_key=api_key)
 
114
  contents=contents,
115
  config=generate_content_config,
116
  ):
117
+ if not chunk.candidates or not chunk.candidates[0].content:
118
  continue
 
119
 
120
+ candidate = chunk.candidates[0].content.parts[0]
121
  if candidate.inline_data:
122
  save_binary_file(temp_path, candidate.inline_data.data)
123
+ print(f"[DEBUG] Gemini returned image -> {temp_path}")
124
  image_path = temp_path
125
  break
126
  else:
 
130
  return image_path, text_response
131
 
132
  #######################################
133
+ # 3. Diffusion (Flux)์šฉ ํ•จ์ˆ˜
134
  #######################################
135
 
136
+ def generate_random_letters(length: int) -> str:
137
+ """length ๊ธธ์ด๋งŒํผ ๋Œ€์†Œ๋ฌธ์ž ์•ŒํŒŒ๋ฒณ์„ ๋ฌด์ž‘์œ„๋กœ ์ƒ์„ฑ."""
138
+ letters = string.ascii_lowercase + string.ascii_uppercase
139
+ return "".join(random.choice(letters) for _ in range(length))
140
+
141
+ def fill_prompt_with_random_texts(prompt: str, r1: str, r2: str, r3: str) -> str:
142
  """
143
+ ํ”„๋กฌํ”„ํŠธ ๋‚ด <text1>, <text2>, <text3>๋ฅผ
144
+ ๊ฐ๊ฐ r1, r2, r3๋กœ ์น˜ํ™˜.
145
+ - <text1>์€ ํ•„์ˆ˜ (์—†์œผ๋ฉด ์ž๋™์œผ๋กœ ๋’ค์— ๋ถ™์ž„).
146
+ - <text2>, <text3>๋Š” ์žˆ์œผ๋ฉด ์น˜ํ™˜, ์—†์œผ๋ฉด ๋ฌด์‹œ.
147
  """
148
+ # 1) <text1>์€ ํ•„์ˆ˜
149
+ if "<text1>" in prompt:
150
+ prompt = prompt.replace("<text1>", r1)
151
  else:
152
+ # ์ž๋™ ๋ง๋ถ™์ž„
153
+ prompt = f"{prompt} with clear readable text that says '{r1}'"
154
+
155
+ # 2) <text2>, <text3>๋Š” ์„ ํƒ
156
+ if "<text2>" in prompt:
157
+ prompt = prompt.replace("<text2>", r2)
158
+ if "<text3>" in prompt:
159
+ prompt = prompt.replace("<text3>", r3)
160
+
161
+ return prompt
162
 
163
+ def generate_initial_image(prompt, random1, random2, random3, height, width, steps, scale, seed):
164
+ """
165
+ Flux ํŒŒ์ดํ”„๋ผ์ธ์„ ์ด์šฉํ•ด (r1, r2, r3)๊ฐ€ ๋“ค์–ด๊ฐ„ ์ด๋ฏธ์ง€๋ฅผ ์ƒ์„ฑ.
166
+ """
167
+ with torch.inference_mode(), torch.autocast("cuda", dtype=torch.bfloat16), timer("Flux Generation"):
168
  result = pipe(
169
+ prompt=[prompt],
170
  generator=torch.Generator().manual_seed(int(seed)),
171
  num_inference_steps=int(steps),
172
  guidance_scale=float(scale),
 
174
  width=int(width),
175
  max_sequence_length=256
176
  ).images[0]
 
177
  return result
178
 
179
+
180
+ def change_multi_text_in_image(original_image, random1, final1, random2, final2, random3, final3):
181
  """
182
+ Gemini๋ฅผ ํ†ตํ•ด, ์ด๋ฏธ์ง€ ์•ˆ์˜ r1->final1, r2->final2, r3->final3 ์‹์œผ๋กœ ํ…์ŠคํŠธ ๊ต์ฒด.
183
+ - r2, final2 (๋˜๋Š” r3, final3)๊ฐ€ ๋นˆ ๋ฌธ์ž์—ด์ด๋ฉด ํ•ด๋‹น ๊ต์ฒด๋Š” ๊ฑด๋„ˆ๋œ€.
184
  """
185
+ # ๊ต์ฒด ์ง€์‹œ๋ฌธ ๋งŒ๋“ค๊ธฐ
186
+ instructions = []
187
+ if random1 and final1:
188
+ instructions.append(f"Change any text reading '{random1}' in this image to '{final1}'.")
189
+ if random2 and final2:
190
+ instructions.append(f"Change any text reading '{random2}' in this image to '{final2}'.")
191
+ if random3 and final3:
192
+ instructions.append(f"Change any text reading '{random3}' in this image to '{final3}'.")
193
+
194
+ # ๋งŒ์•ฝ ๊ต์ฒด ์ง€์‹œ๋ฌธ์ด ์—†๋‹ค๋ฉด ๊ทธ๋ƒฅ return original_image
195
+ if not instructions:
196
+ print("[WARN] No text changes requested!")
197
+ return original_image
198
+
199
+ full_instruction = " ".join(instructions)
200
  try:
201
+ # ์ž„์‹œ ํŒŒ์ผ์— original_image ์ €์žฅ
202
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
203
  original_path = tmp.name
204
  original_image.save(original_path)
205
 
206
  image_path, text_response = generate_by_google_genai(
207
+ text=full_instruction,
208
  file_name=original_path
209
  )
 
210
  if image_path:
211
  with open(image_path, "rb") as f:
212
  image_data = f.read()
213
+ new_img = Image.open(io.BytesIO(image_data))
214
+ return new_img
215
  else:
216
+ # ์ด๋ฏธ์ง€ ์—†์ด ํ…์ŠคํŠธ๋งŒ ์˜จ ๊ฒฝ์šฐ
217
+ print("[WARN] Gemini returned only text:", text_response)
218
+ return original_image
219
  except Exception as e:
220
  raise gr.Error(f"Error: {e}")
221
 
222
  #######################################
223
+ # 4. ๋ฉ”์ธ ํ”„๋กœ์„ธ์Šค ํ•จ์ˆ˜
 
 
 
 
 
 
 
 
 
 
 
 
 
224
  #######################################
225
 
226
+ def run_process(
227
+ prompt,
228
+ final_text1,
229
+ final_text2,
230
+ final_text3,
231
+ height,
232
+ width,
233
+ steps,
234
+ scale,
235
+ seed
236
+ ):
237
  """
238
+ 1) final_text1(ํ•„์ˆ˜), final_text2, final_text3(์˜ต์…˜) ๊ฐ๊ฐ ๊ธธ์ด์— ๋งž์ถฐ ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ ๋งŒ๋“ค๊ธฐ.
239
+ 2) prompt ๋‚ด <text1>, <text2>, <text3> ์น˜ํ™˜ -> Flux๋กœ 1์ฐจ(๋žœ๋ค) ์ด๋ฏธ์ง€.
240
+ 3) Gemini ํ˜ธ์ถœ -> r1->final_text1, r2->final_text2, r3->final_text3 ๊ต์ฒด -> ์ตœ์ข… ์ด๋ฏธ์ง€.
241
  """
242
+ # (A) ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ
243
+ r1 = generate_random_letters(len(final_text1)) if final_text1 else ""
244
+ r2 = generate_random_letters(len(final_text2)) if final_text2 else ""
245
+ r3 = generate_random_letters(len(final_text3)) if final_text3 else ""
246
+
247
+ # (B) ํ”„๋กฌํ”„ํŠธ ์น˜ํ™˜
248
+ final_prompt = fill_prompt_with_random_texts(prompt, r1, r2, r3)
249
+ print(f"[DEBUG] final_prompt = {final_prompt}")
250
+
251
+ # (C) 1์ฐจ ์ด๋ฏธ์ง€ (๋žœ๋ค ํ…์ŠคํŠธ)
252
+ random_image = generate_initial_image(final_prompt, r1, r2, r3, height, width, steps, scale, seed)
253
+
254
+ # (D) 2์ฐจ ์ด๋ฏธ์ง€ (์‹ค์ œ ํ…์ŠคํŠธ)
255
+ final_image = change_multi_text_in_image(
256
+ random_image,
257
+ r1, final_text1,
258
+ r2, final_text2,
259
+ r3, final_text3
260
+ )
261
 
 
 
 
262
  return [random_image, final_image]
263
 
264
  #######################################
265
+ # 5. Gradio UI
266
  #######################################
267
 
268
+ with gr.Blocks(title="Flux + Google GenAI (Up to 3 Text placeholders)") as demo:
269
  gr.Markdown(
270
  """
271
+ # Flux + Google GenAI: ์ตœ๋Œ€ 3๊ฐœ์˜ `<text>` ๊ต์ฒด
272
+
273
+ ## ์‚ฌ์šฉ ๋ฐฉ๋ฒ•
274
+ 1. ์•„๋ž˜ Prompt์— `<text1>`, `<text2>`, `<text3>`๋ฅผ ์ตœ๋Œ€ 3๊ฐœ๊นŒ์ง€ ๋ฐฐ์น˜ ๊ฐ€๋Šฅ.
275
+ - ์˜ˆ) "A poster with <text1> in large letters, also <text2> in the corner"
276
+ - **<text1>์€ ํ•„์ˆ˜**(์—†์œผ๋ฉด ์ž๋™์œผ๋กœ ๋ฌธ๊ตฌ๊ฐ€ ๋’ค์— ๋ถ™์Œ)
277
+ - <text2>, <text3>๋Š” ๋„ฃ์–ด๋„ ๋˜๊ณ , ์•ˆ ๋„ฃ์–ด๋„ ๋จ.
278
+ 2. "New Text #1" (ํ•„์ˆ˜), "New Text #2", "New Text #3"๋ฅผ ์ž…๋ ฅ.
279
+ - #2, #3๋Š” ๋น„์›Œ ๋‘๋ฉด ํ•ด๋‹น ์ž๋ฆฌ ๊ต์ฒด ์—†์Œ.
280
+ 3. "Generate Images" ๋ฒ„ํŠผ โ†’
281
+ (1) `<text1>`, `<text2>`, `<text3>` ์ž๋ฆฌ์— (๋˜๋Š” ์ž๋™์œผ๋กœ) **๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ** ๋„ฃ์€ 1์ฐจ ์ด๋ฏธ์ง€ ์ƒ์„ฑ
282
+ (2) ์ด์–ด Gemini ๋ชจ๋ธ์„ ํ†ตํ•ด ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ โ†’ ์‹ค์ œ "New Text #1/2/3" ๋ณ€๊ฒฝํ•œ 2์ฐจ ์ด๋ฏธ์ง€
283
+ - **๋‘ ์ด๋ฏธ์ง€**(๋žœ๋ค ํ…์ŠคํŠธ โ†’ ์ตœ์ข… ํ…์ŠคํŠธ)๊ฐ€ ์ˆœ์„œ๋Œ€๋กœ ์ถœ๋ ฅ๋ฉ๋‹ˆ๋‹ค.
284
+
285
+ ---
286
  """
287
  )
288
 
289
+ # ์˜ˆ์‹œ 5๊ฐœ
290
+ examples = [
291
+ [
292
+ "A futuristic billboard shows <text1> and a small sign <text2> on the left side. <text3> is a hidden watermark.",
293
+ "HELLO", "WELCOME", "2025"
294
+ ],
295
+ [
296
+ "A fantasy poster with <text1> and <text2> in stylized letters, plus a tiny note <text3> at the bottom.",
297
+ "Dragons", "MagicRealm", "Beware!"
298
+ ],
299
+ [
300
+ "A neon sign reading <text1>, with a secondary text <text2> below. <text3> might appear in the corner.",
301
+ "OPEN", "24HOUR", "NoSmoking"
302
+ ],
303
+ [
304
+ "A big invitation card with main text <text1>, subtitle <text2>, signature <text3> in cursive.",
305
+ "Birthday Party", "Today Only", "From Your Friend"
306
+ ],
307
+ [
308
+ "A large graffiti wall with <text1> in bold letters, plus <text2> and <text3> near the edges.",
309
+ "FREEDOM", "HOPE", "LOVE"
310
+ ]
311
+ ]
312
+
313
  with gr.Row():
314
  with gr.Column():
315
  prompt_input = gr.Textbox(
316
  lines=3,
317
+ label="Prompt (use `<text1>`, `<text2>`, `<text3>` as needed)",
318
+ placeholder="Ex) A poster with <text1>, plus a line <text2>, etc."
319
+ )
320
+ final_text1 = gr.Textbox(
321
+ label="New Text #1 (Required)",
322
+ placeholder="Ex) HELLO"
323
  )
324
+ final_text2 = gr.Textbox(
325
+ label="New Text #2 (Optional)",
326
+ placeholder="Ex) WELCOME"
 
327
  )
328
+ final_text3 = gr.Textbox(
329
+ label="New Text #3 (Optional)",
330
+ placeholder="Ex) 2025 or anything"
331
+ )
332
+
333
+ with gr.Accordion("Advanced Settings", open=False):
334
  height = gr.Slider(label="Height", minimum=256, maximum=1152, step=64, value=512)
335
  width = gr.Slider(label="Width", minimum=256, maximum=1152, step=64, value=512)
336
  steps = gr.Slider(label="Inference Steps", minimum=6, maximum=25, step=1, value=8)
 
339
 
340
  run_btn = gr.Button("Generate Images", variant="primary")
341
 
342
+ gr.Examples(
343
+ examples=examples,
344
+ inputs=[prompt_input, final_text1, final_text2, final_text3],
345
+ label="Click to load example"
346
+ )
347
+
348
  with gr.Column():
349
  random_image_output = gr.Image(label="1) Random Text Image", type="pil")
350
  final_image_output = gr.Image(label="2) Final Text Image", type="pil")
351
 
352
+ # ๋ฒ„ํŠผ ์•ก์…˜
353
  run_btn.click(
354
+ fn=run_process,
355
+ inputs=[
356
+ prompt_input,
357
+ final_text1,
358
+ final_text2,
359
+ final_text3,
360
+ height,
361
+ width,
362
+ steps,
363
+ scale,
364
+ seed
365
+ ],
366
  outputs=[random_image_output, final_image_output]
367
  )
368