ginipick commited on
Commit
819fc44
ยท
verified ยท
1 Parent(s): 15cc8b5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +78 -108
app.py CHANGED
@@ -35,7 +35,7 @@ os.environ["TRANSFORMERS_CACHE"] = CACHE_PATH
35
  os.environ["HF_HUB_CACHE"] = CACHE_PATH
36
  os.environ["HF_HOME"] = CACHE_PATH
37
 
38
- # ๊ฐ„๋‹จํ•œ ํƒ€์ด๋จธ ํด๋ž˜์Šค
39
  class timer:
40
  def __init__(self, method_name="timed process"):
41
  self.method = method_name
@@ -65,7 +65,7 @@ pipe.fuse_lora(lora_scale=0.125)
65
  pipe.to(device="cuda", dtype=torch.bfloat16)
66
 
67
  #######################################
68
- # 2. Google GenAI๋ฅผ ํ†ตํ•œ ์ด๋ฏธ์ง€ ๋‚ด ํ…์ŠคํŠธ ๋ณ€ํ™˜ ํ•จ์ˆ˜
69
  #######################################
70
 
71
  def save_binary_file(file_name, data):
@@ -80,21 +80,16 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
80
  - file_name: ์›๋ณธ ์ด๋ฏธ์ง€(์˜ˆ: .png) ๊ฒฝ๋กœ
81
  - model: ์‚ฌ์šฉํ•  gemini ๋ชจ๋ธ ์ด๋ฆ„
82
  """
83
- # (1) ํ™˜๊ฒฝ ๋ณ€์ˆ˜์—์„œ API ํ‚ค ๊ฐ€์ ธ์˜ค๊ธฐ (ํ•„์ˆ˜)
84
  api_key = os.getenv("GAPI_TOKEN", None)
85
  if not api_key:
86
  raise ValueError(
87
  "GAPI_TOKEN ํ™˜๊ฒฝ ๋ณ€์ˆ˜๊ฐ€ ์„ค์ •๋˜์ง€ ์•Š์•˜์Šต๋‹ˆ๋‹ค. "
88
- "Google GenAI API๋ฅผ ์‚ฌ์šฉํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” GAPI_TOKEN์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค."
89
  )
90
 
91
- # (2) Google Client ์ดˆ๊ธฐํ™”
92
  client = genai.Client(api_key=api_key)
93
-
94
- # (3) ์ด๋ฏธ์ง€ ์—…๋กœ๋“œ
95
  files = [client.files.upload(file=file_name)]
96
 
97
- # (4) gemini์— ์ „๋‹ฌํ•  Content ์ค€๋น„ (์ด๋ฏธ์ง€ + ํ”„๋กฌํ”„ํŠธ)
98
  contents = [
99
  types.Content(
100
  role="user",
@@ -108,7 +103,6 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
108
  ),
109
  ]
110
 
111
- # (5) ์ƒ์„ฑ/๋ณ€ํ™˜ ์„ค์ •
112
  generate_content_config = types.GenerateContentConfig(
113
  temperature=1,
114
  top_p=0.95,
@@ -121,10 +115,8 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
121
  text_response = ""
122
  image_path = None
123
 
124
- # ์ž„์‹œ ํŒŒ์ผ๋กœ ์ด๋ฏธ์ง€ ๋ฐ›์„ ์ค€๋น„
125
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
126
  temp_path = tmp.name
127
- # ์‘๋‹ต ์ŠคํŠธ๋ฆผ ์ฒ˜๋ฆฌ
128
  for chunk in client.models.generate_content_stream(
129
  model=model,
130
  contents=contents,
@@ -134,60 +126,34 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
134
  continue
135
  candidate = chunk.candidates[0].content.parts[0]
136
 
137
- # inline_data๊ฐ€ ์žˆ์œผ๋ฉด ์ด๋ฏธ์ง€ ์‘๋‹ต
138
  if candidate.inline_data:
139
  save_binary_file(temp_path, candidate.inline_data.data)
140
  print(f"File of mime type {candidate.inline_data.mime_type} saved to: {temp_path}")
141
  image_path = temp_path
142
  break
143
  else:
144
- # ์ด๋ฏธ์ง€ ์—†์ด ํ…์ŠคํŠธ๋งŒ ๋ฐ˜ํ™˜๋˜๋Š” ๊ฒฝ์šฐ
145
  text_response += chunk.text + "\n"
146
 
147
- # ์—…๋กœ๋“œํ•œ File ๊ฐ์ฒด ์ œ๊ฑฐ
148
  del files
149
-
150
  return image_path, text_response
151
 
152
  #######################################
153
- # 3. ์ž„์˜ ์•ŒํŒŒ๋ฒณ ์ƒ์„ฑ์šฉ ํ•จ์ˆ˜
154
- #######################################
155
-
156
- def generate_random_letters(length: int) -> str:
157
- """
158
- length ๊ธธ์ด์˜ ์ž„์˜์˜ ์•ŒํŒŒ๋ฒณ ๋ฌธ์ž์—ด์„ ์ƒ์„ฑํ•ด ๋ฐ˜ํ™˜.
159
- ๋Œ€๋ฌธ์ž+์†Œ๋ฌธ์ž(a-z, A-Z) ๋ฒ”์œ„์—์„œ ๋žœ๋ค ๋ฝ‘๊ธฐ.
160
- """
161
- letters = string.ascii_uppercase + string.ascii_lowercase
162
- return "".join(random.choice(letters) for _ in range(length))
163
-
164
- def fill_text_input_with_random(new_text: str) -> str:
165
- """
166
- "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"์˜ ๊ธธ์ด๋ฅผ ์„ธ์–ด, ๊ทธ ๊ธธ์ด๋งŒํผ ์ž„์˜์˜ ์•ŒํŒŒ๋ฒณ ๋ฌธ์ž์—ด ์ƒ์„ฑ ํ›„ ๋ฐ˜ํ™˜.
167
- ์ด ๋ฐ˜ํ™˜๊ฐ’์ด ๊ณง "์ด๋ฏธ์ง€ ์•ˆ์— ๋“ค์–ด๊ฐˆ ํ…์ŠคํŠธ" UI(=text_input)์— ํ‘œ์‹œ๋จ.
168
- """
169
- length = len(new_text)
170
- random_letters = generate_random_letters(length)
171
- return random_letters
172
-
173
- #######################################
174
- # 4. Gradio ํ•จ์ˆ˜
175
  #######################################
176
 
177
  def generate_initial_image(prompt, text, height, width, steps, scale, seed):
178
  """
179
- FLUX ํŒŒ์ดํ”„๋ผ์ธ์„ ์‚ฌ์šฉํ•ด 'ํ…์ŠคํŠธ๊ฐ€ ํฌํ•จ๋œ ์ด๋ฏธ์ง€๋ฅผ' ๋จผ์ € ์ƒ์„ฑ.
180
- - prompt ๋‚ด <text>๋ฅผ text๋กœ ์น˜ํ™˜
181
- - <text>๊ฐ€ ์—†๋‹ค๋ฉด "with clear readable text that says '<text>'"๋ฅผ ์ž๋™ ๋ถ™์ž„
182
  """
183
  if "<text>" in prompt:
184
  combined_prompt = prompt.replace("<text>", text)
185
  else:
186
  combined_prompt = f"{prompt} with clear readable text that says '{text}'"
187
 
188
- print(f"[DEBUG] Final combined_prompt: {combined_prompt}")
189
 
190
- with torch.inference_mode(), torch.autocast("cuda", dtype=torch.bfloat16), timer("inference"):
191
  result = pipe(
192
  prompt=[combined_prompt],
193
  generator=torch.Generator().manual_seed(int(seed)),
@@ -202,9 +168,8 @@ def generate_initial_image(prompt, text, height, width, steps, scale, seed):
202
 
203
  def change_text_in_image(original_image, new_text):
204
  """
205
- Google GenAI์˜ gemini ๋ชจ๋ธ์„ ํ†ตํ•ด,
206
- ์—…๋กœ๋“œ๋œ ์ด๋ฏธ์ง€ ๋‚ด๋ถ€์˜ ๋ฌธ๊ตฌ๋ฅผ `new_text`๋กœ ๋ณ€๊ฒฝํ•ด์ฃผ๋Š” ํ•จ์ˆ˜.
207
- (์—ฌ๊ธฐ์„œ๋Š” new_text๊ฐ€ ์ด๋ฏธ ์ž„์˜์˜ ์•ŒํŒŒ๋ฒณ์œผ๋กœ ์ฑ„์›Œ์ง„ ์ƒํƒœ๊ฐ€ ๋จ)
208
  """
209
  try:
210
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
@@ -220,44 +185,84 @@ def change_text_in_image(original_image, new_text):
220
  with open(image_path, "rb") as f:
221
  image_data = f.read()
222
  modified_img = Image.open(io.BytesIO(image_data))
223
- return modified_img, ""
224
  else:
225
- return None, text_response
 
226
 
227
  except Exception as e:
228
  raise gr.Error(f"Error: {e}")
229
 
230
  #######################################
231
- # 5. Gradio ์ธํ„ฐํŽ˜์ด์Šค ๊ตฌ์„ฑ
232
  #######################################
233
 
234
- with gr.Blocks(title="Flux + Google GenAI + Random Text Replacement") as demo:
235
- gr.Markdown(
236
- """
237
- # Flux ๊ธฐ๋ฐ˜ ์ด๋ฏธ์ง€ ์ƒ์„ฑ + Google GenAI + ์ž„์˜ ์•ŒํŒŒ๋ฒณ ์น˜ํ™˜
 
 
238
 
239
- **๊ธฐ๋Šฅ ์š”์•ฝ**
240
- 1) "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"์— ์ž…๋ ฅ๋œ ๋ฌธ์ž์—ด ๊ธธ์ด๋ฅผ ์ธก์ •
241
- 2) ๊ทธ ๊ธธ์ด๋งŒํผ **๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ**(๋Œ€์†Œ๋ฌธ์ž)์œผ๋กœ ๊ตฌ์„ฑ๋œ ๋ฌธ์ž์—ด์„ ๋งŒ๋“ค์–ด
242
- 3) "์ด๋ฏธ์ง€ ์•ˆ์— ๋“ค์–ด๊ฐˆ ํ…์ŠคํŠธ" ์นธ์— ์ž๋™์œผ๋กœ ์ž…๋ ฅ (์ฆ‰, ์‹ค์ œ ๋ฐ”๋€” ํ…์ŠคํŠธ๋Š” ์™„์ „ํžˆ ๋žœ๋ค)
243
- 4) ๊ทธ ํ›„ Google GenAI(gemini)๋กœ ์ด๋ฏธ์ง€ ๋ฌธ์ž๋ฅผ ๊ต์ฒด
 
 
 
 
 
 
 
 
 
 
244
 
245
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
246
  """
247
  )
248
 
249
  with gr.Row():
250
  with gr.Column():
251
- gr.Markdown("## 1) Step 1: FLUX๋กœ ํ…์ŠคํŠธ ํฌํ•จ ์ด๋ฏธ์ง€ ์ƒ์„ฑ")
252
  prompt_input = gr.Textbox(
253
  lines=3,
254
- label="์ด๋ฏธ์ง€ ์žฅ๋ฉด/๋ฐฐ๊ฒฝ Prompt (use `<text>` placeholder if you like)",
255
- placeholder="e.g. A white cat with speech bubble says <text>"
256
  )
257
- text_input = gr.Textbox(
258
  lines=1,
259
- label="์ด๋ฏธ์ง€ ์•ˆ์— ๋“ค์–ด๊ฐˆ ํ…์ŠคํŠธ",
260
- placeholder="e.g. Hello or ์•ˆ๋…•"
261
  )
262
  with gr.Accordion("๊ณ ๊ธ‰ ์„ค์ • (ํ™•์žฅ)", open=False):
263
  height = gr.Slider(label="Height", minimum=256, maximum=1152, step=64, value=512)
@@ -266,52 +271,17 @@ with gr.Blocks(title="Flux + Google GenAI + Random Text Replacement") as demo:
266
  scale = gr.Slider(label="Guidance Scale", minimum=0.0, maximum=10.0, step=0.5, value=3.5)
267
  seed = gr.Number(label="Seed (reproducibility)", value=1234, precision=0)
268
 
269
- generate_btn = gr.Button("Generate Base Image", variant="primary")
270
- generated_image = gr.Image(label="Generated Image (with text)", type="pil")
271
 
272
  with gr.Column():
273
- gr.Markdown(
274
- """
275
- ## 2) Step 2: "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ" ๊ธธ์ด๋ฅผ ๊ธฐ์ค€์œผ๋กœ
276
- ์ž„์˜ ์•ŒํŒŒ๋ฒณ์„ "์ด๋ฏธ์ง€ ์•ˆ์— ๋“ค์–ด๊ฐˆ ํ…์ŠคํŠธ"๋กœ ์ž๋™ ์„ค์ • โ†’ ์ด๋ฏธ์ง€ ๊ต์ฒด
277
- """
278
- )
279
- new_text_input = gr.Textbox(
280
- label="์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ (์ž…๋ ฅ๋œ ๊ธธ์ด๋ฅผ ๊ธฐ์ค€์œผ๋กœ ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ ์ƒ์„ฑ)",
281
- placeholder="์˜ˆ) Hello123 (๊ธธ์ด 8์ด๋ฉด, 8๊ธ€์ž ๋žœ๋ค)"
282
- )
283
-
284
- # (A) ๋จผ์ € "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ" ๊ธ€์ž์ˆ˜๋ฅผ ์„ธ์–ด, ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ์œผ๋กœ text_input ์ฑ„์šฐ๊ธฐ
285
- # (B) ๊ทธ๋Ÿฐ ๋’ค change_text_in_image() ์‹คํ–‰
286
- modify_btn = gr.Button("Generate Random Letters & Change Text in Image", variant="secondary")
287
-
288
- # ์ตœ์ข… ๊ฒฐ๊ณผ
289
- output_img = gr.Image(label="Modified Image", type="pil")
290
- output_txt = gr.Textbox(label="(If only text returned)")
291
-
292
- ###########################
293
- # 1) Step1 ๋ฒ„ํŠผ ์•ก์…˜
294
- ###########################
295
- generate_btn.click(
296
- fn=generate_initial_image,
297
- inputs=[prompt_input, text_input, height, width, steps, scale, seed],
298
- outputs=[generated_image]
299
- )
300
-
301
- ###########################
302
- # 2) Step2 ๋ฒ„ํŠผ ์•ก์…˜
303
- ###########################
304
- # (A) ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ ์ƒ์„ฑํ•ด text_input์— ๋ฐ˜์˜
305
- chain = modify_btn.click(
306
- fn=fill_text_input_with_random,
307
- inputs=[new_text_input],
308
- outputs=[text_input]
309
- )
310
- # (B) ๋ฐฉ๊ธˆ ์—…๋ฐ์ดํŠธ๋œ text_input ๊ฐ’์„ ์ด์šฉํ•ด, ์ด๋ฏธ์ง€ ํ…์ŠคํŠธ ๊ต์ฒด
311
- chain.then(
312
- fn=change_text_in_image,
313
- inputs=[generated_image, text_input],
314
- outputs=[output_img, output_txt]
315
  )
316
 
317
  demo.launch(max_threads=20)
 
35
  os.environ["HF_HUB_CACHE"] = CACHE_PATH
36
  os.environ["HF_HOME"] = CACHE_PATH
37
 
38
+ # ํƒ€์ด๋จธ ํด๋ž˜์Šค
39
  class timer:
40
  def __init__(self, method_name="timed process"):
41
  self.method = method_name
 
65
  pipe.to(device="cuda", dtype=torch.bfloat16)
66
 
67
  #######################################
68
+ # 2. Google GenAI ๋ชจ๋ธ๋กœ ํ…์ŠคํŠธ ๋ณ€ํ™˜ ํ•จ์ˆ˜
69
  #######################################
70
 
71
  def save_binary_file(file_name, data):
 
80
  - file_name: ์›๋ณธ ์ด๋ฏธ์ง€(์˜ˆ: .png) ๊ฒฝ๋กœ
81
  - model: ์‚ฌ์šฉํ•  gemini ๋ชจ๋ธ ์ด๋ฆ„
82
  """
 
83
  api_key = os.getenv("GAPI_TOKEN", None)
84
  if not api_key:
85
  raise ValueError(
86
  "GAPI_TOKEN ํ™˜๊ฒฝ ๋ณ€์ˆ˜๊ฐ€ ์„ค์ •๋˜์ง€ ์•Š์•˜์Šต๋‹ˆ๋‹ค. "
87
+ "Google GenAI API ์‚ฌ์šฉ์„ ์œ„ํ•ด์„œ๋Š” GAPI_TOKEN์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค."
88
  )
89
 
 
90
  client = genai.Client(api_key=api_key)
 
 
91
  files = [client.files.upload(file=file_name)]
92
 
 
93
  contents = [
94
  types.Content(
95
  role="user",
 
103
  ),
104
  ]
105
 
 
106
  generate_content_config = types.GenerateContentConfig(
107
  temperature=1,
108
  top_p=0.95,
 
115
  text_response = ""
116
  image_path = None
117
 
 
118
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
119
  temp_path = tmp.name
 
120
  for chunk in client.models.generate_content_stream(
121
  model=model,
122
  contents=contents,
 
126
  continue
127
  candidate = chunk.candidates[0].content.parts[0]
128
 
 
129
  if candidate.inline_data:
130
  save_binary_file(temp_path, candidate.inline_data.data)
131
  print(f"File of mime type {candidate.inline_data.mime_type} saved to: {temp_path}")
132
  image_path = temp_path
133
  break
134
  else:
 
135
  text_response += chunk.text + "\n"
136
 
 
137
  del files
 
138
  return image_path, text_response
139
 
140
  #######################################
141
+ # 3. Diffusion + GoogleGenAI๋ฅผ ์—ฐ๊ฒฐ
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142
  #######################################
143
 
144
  def generate_initial_image(prompt, text, height, width, steps, scale, seed):
145
  """
146
+ 1) FLUX ํŒŒ์ดํ”„๋ผ์ธ์„ ์‚ฌ์šฉํ•ด 'text'๊ฐ€ ๋“ค์–ด๊ฐ„ ์ด๋ฏธ์ง€๋ฅผ ์ƒ์„ฑ
147
+ - prompt ๋‚ด <text>๊ฐ€ ์žˆ์œผ๋ฉด ์น˜ํ™˜, ์—†์œผ๋ฉด ์ž๋™ ์ถ”๊ฐ€
 
148
  """
149
  if "<text>" in prompt:
150
  combined_prompt = prompt.replace("<text>", text)
151
  else:
152
  combined_prompt = f"{prompt} with clear readable text that says '{text}'"
153
 
154
+ print("[DEBUG] combined_prompt:", combined_prompt)
155
 
156
+ with torch.inference_mode(), torch.autocast("cuda", dtype=torch.bfloat16), timer("GenerateInitialImage"):
157
  result = pipe(
158
  prompt=[combined_prompt],
159
  generator=torch.Generator().manual_seed(int(seed)),
 
168
 
169
  def change_text_in_image(original_image, new_text):
170
  """
171
+ 2) Gemini ๋ชจ๋ธ์„ ํ†ตํ•ด,
172
+ original_image ๋‚ด ํ…์ŠคํŠธ๋ฅผ `new_text`๋กœ ๋ณ€๊ฒฝํ•œ ์ด๋ฏธ์ง€ ๋ฐ˜ํ™˜
 
173
  """
174
  try:
175
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
 
185
  with open(image_path, "rb") as f:
186
  image_data = f.read()
187
  modified_img = Image.open(io.BytesIO(image_data))
188
+ return modified_img
189
  else:
190
+ # ์ด๋ฏธ์ง€๊ฐ€ ์—†์ด ํ…์ŠคํŠธ๋งŒ ์‘๋‹ต๋œ ๊ฒฝ์šฐ
191
+ return None
192
 
193
  except Exception as e:
194
  raise gr.Error(f"Error: {e}")
195
 
196
  #######################################
197
+ # 4. ์ž„์˜ ์•ŒํŒŒ๋ฒณ ์ƒ์„ฑ
198
  #######################################
199
 
200
+ def generate_random_letters(length: int) -> str:
201
+ """
202
+ length ๊ธธ์ด์˜ ์ž„์˜ ์•ŒํŒŒ๋ฒณ(๋Œ€์†Œ๋ฌธ์ž) ๋ฌธ์ž์—ด ์ƒ์„ฑ
203
+ """
204
+ letters = string.ascii_lowercase + string.ascii_uppercase
205
+ return "".join(random.choice(letters) for _ in range(length))
206
 
207
+ #######################################
208
+ # 5. ์ตœ์ข… ํ•จ์ˆ˜: ๋ฒ„ํŠผ ํ•œ ๋ฒˆ์œผ๋กœ
209
+ # (1) ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ์œผ๋กœ 1์ฐจ ์ด๋ฏธ์ง€ ์ƒ์„ฑ
210
+ # (2) ์ง„์งœ "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"๋กœ 2์ฐจ ์ด๋ฏธ์ง€ ์ƒ์„ฑ
211
+ #######################################
212
+
213
+ def run_full_process(prompt, final_text, height, width, steps, scale, seed):
214
+ """
215
+ - final_text์˜ ๊ธธ์ด์— ๋งž์ถฐ ๋žœ๋ค ์•ŒํŒŒ๋ฒณ์„ ์ƒ์„ฑ -> 1์ฐจ ์ด๋ฏธ์ง€
216
+ - ๊ทธ 1์ฐจ ์ด๋ฏธ์ง€๋ฅผ ๋ฐ”ํƒ•์œผ๋กœ, final_text๋กœ ๊ต์ฒด -> 2์ฐจ ์ตœ์ข… ์ด๋ฏธ์ง€
217
+ """
218
+ # (A) ์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ(final_text) ๊ธ€์ž์ˆ˜๋งŒํผ ์ž„์˜ ์•ŒํŒŒ๋ฒณ ์ƒ์„ฑ
219
+ random_len = len(final_text)
220
+ random_text = generate_random_letters(random_len)
221
+ print(f"[STEP] final_text='{final_text}' => random_text='{random_text}'")
222
 
223
+ # (B) 1์ฐจ ์ด๋ฏธ์ง€: ๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ์œผ๋กœ ์ƒ์„ฑ
224
+ random_image = generate_initial_image(prompt, random_text, height, width, steps, scale, seed)
225
+
226
+ # (C) 2์ฐจ ์ด๋ฏธ์ง€: ์‹ค์ œ final_text๋กœ ๊ต์ฒด
227
+ final_image = change_text_in_image(random_image, final_text)
228
+
229
+ return [random_image, final_image]
230
+
231
+ #######################################
232
+ # 6. Gradio UI
233
+ #######################################
234
+
235
+ with gr.Blocks(title="Flux + Google GenAI (Random & Then Real Text)") as demo:
236
+ gr.Markdown(
237
+ """
238
+ # Flux + Google GenAI: ๋‘ ๋‹จ๊ณ„์— ๊ฑธ์นœ ํ…์ŠคํŠธ ๊ต์ฒด
239
+
240
+ **์‚ฌ์šฉ ํ๋ฆ„**
241
+ 1) Prompt์— ์žฅ๋ฉด์ด๋‚˜ ์Šคํƒ€์ผ์„ ์ž‘์„ฑ (ํ•„์š”ํ•˜๋ฉด `<text>` ๊ตฌ๋ถ„์ž ์‚ฌ์šฉ)
242
+ 2) "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ" ์— ์ตœ์ข… ์›ํ•˜๋Š” ๋ฌธ์ž์—ด์„ ์ž…๋ ฅ (์˜ˆ: "์•ˆ๋…•ํ•˜์„ธ์š”")
243
+ 3) "Generate Images" ๋ฒ„ํŠผ์„ ๋ˆ„๋ฅด๋ฉด,
244
+ - ๋จผ์ € "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ" ๊ธธ์ด์— ๋งž๋Š” **๋ฌด์ž‘์œ„ ์•ŒํŒŒ๋ฒณ**์„ ๋„ฃ์–ด ์ด๋ฏธ์ง€ ์ƒ์„ฑ (1์ฐจ ์ด๋ฏธ์ง€)
245
+ - ์ด์–ด์„œ **์ง„์งœ** "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"๋กœ ๋‹ค์‹œ ๊ต์ฒด(2์ฐจ ์ตœ์ข… ์ด๋ฏธ์ง€)
246
+ 4) ๊ฒฐ๊ณผ๋กœ ๋‘ ์žฅ์˜ ์ด๋ฏธ์ง€๋ฅผ ํ™•์ธํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
247
+
248
+ **์ฃผ์š” ํฌ์ธํŠธ**
249
+ - "์ด๋ฏธ์ง€ ์•ˆ์— ๋“ค์–ด๊ฐˆ ํ…์ŠคํŠธ"๋Š” **UI์— ๋…ธ์ถœ๋˜์ง€ ์•Š์œผ๋ฉฐ**(์‚ฌ์šฉ์ž ์ž…๋ ฅ ๋ถˆ๊ฐ€), ์˜ค์ง ๋‚ด๋ถ€์—์„œ ์ž๋™ ์„ค์ •๋ฉ๋‹ˆ๋‹ค.
250
+ - 1์ฐจ ์ด๋ฏธ์ง€๋Š” ์™„์ „ํžˆ ์ž„์˜์˜ ์•ŒํŒŒ๋ฒณ ํ…์ŠคํŠธ๋ฅผ ํฌํ•จํ•ฉ๋‹ˆ๋‹ค.
251
+ - 2์ฐจ ์ด๋ฏธ์ง€๋Š” ์ตœ์ข…์ ์œผ๋กœ ์‚ฌ์šฉ์ž๊ฐ€ ์ž…๋ ฅํ•œ "์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ"๋ฅผ ํฌํ•จํ•ฉ๋‹ˆ๋‹ค.
252
  """
253
  )
254
 
255
  with gr.Row():
256
  with gr.Column():
 
257
  prompt_input = gr.Textbox(
258
  lines=3,
259
+ label="Prompt (use `<text>` if you want)",
260
+ placeholder="e.g. A white cat with a speech bubble <text>"
261
  )
262
+ final_text_input = gr.Textbox(
263
  lines=1,
264
+ label="์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ",
265
+ placeholder="์˜ˆ) ์•ˆ๋…•ํ•˜์„ธ์š”"
266
  )
267
  with gr.Accordion("๊ณ ๊ธ‰ ์„ค์ • (ํ™•์žฅ)", open=False):
268
  height = gr.Slider(label="Height", minimum=256, maximum=1152, step=64, value=512)
 
271
  scale = gr.Slider(label="Guidance Scale", minimum=0.0, maximum=10.0, step=0.5, value=3.5)
272
  seed = gr.Number(label="Seed (reproducibility)", value=1234, precision=0)
273
 
274
+ run_btn = gr.Button("Generate Images", variant="primary")
 
275
 
276
  with gr.Column():
277
+ random_image_output = gr.Image(label="1) Random Text Image", type="pil")
278
+ final_image_output = gr.Image(label="2) Final Text Image", type="pil")
279
+
280
+ # ๋ฒ„ํŠผ ์•ก์…˜: ์œ„ ๋‹จ๊ณ„๋“ค์„ ๋ชจ๋‘ ์‹คํ–‰ -> ๊ฒฐ๊ณผ 2์žฅ ์ถœ๋ ฅ
281
+ run_btn.click(
282
+ fn=run_full_process,
283
+ inputs=[prompt_input, final_text_input, height, width, steps, scale, seed],
284
+ outputs=[random_image_output, final_image_output]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
285
  )
286
 
287
  demo.launch(max_threads=20)