ginipick commited on
Commit
f8cac3e
ยท
verified ยท
1 Parent(s): b2f5030

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +43 -35
app.py CHANGED
@@ -78,7 +78,7 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
78
  - file_name: ์›๋ณธ ์ด๋ฏธ์ง€(์˜ˆ: .png) ๊ฒฝ๋กœ
79
  - model: ์‚ฌ์šฉํ•  gemini ๋ชจ๋ธ ์ด๋ฆ„
80
  """
81
- # GAPI_TOKEN ํ™˜๊ฒฝ๋ณ€์ˆ˜์—์„œ ํ‚ค๋ฅผ ๊ฐ€์ ธ์˜ด (ํ•„์ˆ˜)
82
  api_key = os.getenv("GAPI_TOKEN", None)
83
  if not api_key:
84
  raise ValueError(
@@ -86,12 +86,13 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
86
  "Google GenAI API๋ฅผ ์‚ฌ์šฉํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” GAPI_TOKEN์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค."
87
  )
88
 
 
89
  client = genai.Client(api_key=api_key)
90
 
91
- # ์ด๋ฏธ์ง€ ์—…๋กœ๋“œ
92
  files = [client.files.upload(file=file_name)]
93
 
94
- # gemini์— ์ „๋‹ฌํ•  Content ์ค€๋น„
95
  contents = [
96
  types.Content(
97
  role="user",
@@ -105,6 +106,7 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
105
  ),
106
  ]
107
 
 
108
  generate_content_config = types.GenerateContentConfig(
109
  temperature=1,
110
  top_p=0.95,
@@ -117,9 +119,10 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
117
  text_response = ""
118
  image_path = None
119
 
120
- # ์ž„์‹œ ํŒŒ์ผ์— ์ด๋ฏธ์ง€ ์‘๋‹ต์„ ์ €์žฅํ•  ์ค€๋น„
121
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
122
  temp_path = tmp.name
 
123
  for chunk in client.models.generate_content_stream(
124
  model=model,
125
  contents=contents,
@@ -129,32 +132,39 @@ def generate_by_google_genai(text, file_name, model="gemini-2.0-flash-exp"):
129
  continue
130
  candidate = chunk.candidates[0].content.parts[0]
131
 
132
- # inline_data(์ด๋ฏธ์ง€) ์‘๋‹ต์ธ ๊ฒฝ์šฐ
133
  if candidate.inline_data:
134
  save_binary_file(temp_path, candidate.inline_data.data)
135
  print(f"File of mime type {candidate.inline_data.mime_type} saved to: {temp_path}")
136
  image_path = temp_path
137
  break
138
  else:
 
139
  text_response += chunk.text + "\n"
140
 
 
141
  del files
 
142
  return image_path, text_response
143
 
144
  #######################################
145
- # 3. Gradio ํ•จ์ˆ˜
 
146
  #######################################
147
 
148
  def generate_initial_image(prompt, text, height, width, steps, scale, seed):
149
  """
150
- FLUX๋ฅผ ์ด์šฉํ•ด ํ…์ŠคํŠธ๊ฐ€ ํฌํ•จ๋œ ์ด๋ฏธ์ง€๋ฅผ ์ƒ์„ฑ
151
- - prompt ๋‚ด์— <text>๋ผ๋Š” ํŠน์ˆ˜ ๊ตฌ๋ถ„์ž๊ฐ€ ์žˆ์œผ๋ฉด, ๊ฑฐ๊ธฐ์— text๊ฐ€ ์น˜ํ™˜๋จ.
152
- - ๊ทธ๋ ‡์ง€ ์•Š์€ ๊ฒฝ์šฐ, ๊ธฐ์กด์ฒ˜๋Ÿผ prompt ๋’ค์— โ€œwith clear readable text that says ...โ€๋ฅผ ์ถ”๊ฐ€.
153
  """
154
  if "<text>" in prompt:
155
  combined_prompt = prompt.replace("<text>", text)
156
  else:
157
  combined_prompt = f"{prompt} with clear readable text that says '{text}'"
 
 
 
158
 
159
  with torch.inference_mode(), torch.autocast("cuda", dtype=torch.bfloat16), timer("inference"):
160
  result = pipe(
@@ -171,11 +181,10 @@ def generate_initial_image(prompt, text, height, width, steps, scale, seed):
171
 
172
  def change_text_in_image(original_image, new_text):
173
  """
174
- Gemini ๋ชจ๋ธ์„ ํ†ตํ•ด,
175
  ์—…๋กœ๋“œ๋œ ์ด๋ฏธ์ง€ ๋‚ด๋ถ€์˜ ๋ฌธ๊ตฌ๋ฅผ `new_text`๋กœ ๋ณ€๊ฒฝํ•ด์ฃผ๋Š” ํ•จ์ˆ˜.
176
  """
177
  try:
178
- # ์ž„์‹œ ํŒŒ์ผ์— ๋จผ์ € ์ €์žฅ
179
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
180
  original_path = tmp.name
181
  original_image.save(original_path)
@@ -187,35 +196,34 @@ def change_text_in_image(original_image, new_text):
187
  )
188
 
189
  if image_path:
190
- # Gradio ๊ตฌ๋ฒ„์ „์—๋Š” decode_base64_to_image๊ฐ€ ์—†์œผ๋ฏ€๋กœ PIL์„ ์ง์ ‘ ์‚ฌ์šฉ
191
  with open(image_path, "rb") as f:
192
  image_data = f.read()
193
  modified_img = Image.open(io.BytesIO(image_data))
194
  return modified_img, ""
195
  else:
 
196
  return None, text_response
197
 
198
  except Exception as e:
199
  raise gr.Error(f"Error: {e}")
200
 
201
  #######################################
202
- # 4. Gradio ์ธํ„ฐํŽ˜์ด์Šค
203
  #######################################
204
 
205
  with gr.Blocks(title="Flux + Google GenAI Text Replacement") as demo:
206
  gr.Markdown(
207
  """
208
- # Flux Image Generation + Google GenAI Text Replacement
209
-
210
- **Usage Instructions (in English)**
211
- 1. Write a prompt that may contain the special placeholder `<text>`.
212
- - Example: `A white cat says <text> in a cartoon style`.
213
- 2. Enter the actual text in the "Text to Include in the Image" field.
214
- - Example: `์•ˆ๋…•`
215
- 3. Click the "Generate Base Image" button.
216
- - The prompt will be transformed so that `<text>` is replaced with your actual text.
217
- - If `<text>` is **not** found, the text will be appended automatically as `with clear readable text that says ...`.
218
- 4. (Optional) If you want to change the text again, use the "Change Text in Image" button.
219
 
220
  ---
221
  """
@@ -223,32 +231,32 @@ with gr.Blocks(title="Flux + Google GenAI Text Replacement") as demo:
223
 
224
  with gr.Row():
225
  with gr.Column():
226
- gr.Markdown("## 1) Generate the Base Image (FLUX)")
227
  prompt_input = gr.Textbox(
228
  lines=3,
229
- label="Prompt (with optional `<text>` placeholder)",
230
- placeholder="e.g. A white cat says <text> in a cartoon style"
231
  )
232
  text_input = gr.Textbox(
233
  lines=1,
234
- label="Text to Include in the Image",
235
- placeholder="e.g. ์•ˆ๋…•"
236
  )
237
- with gr.Accordion("Advanced Settings", open=False):
238
  height = gr.Slider(label="Height", minimum=256, maximum=1152, step=64, value=512)
239
  width = gr.Slider(label="Width", minimum=256, maximum=1152, step=64, value=512)
240
  steps = gr.Slider(label="Inference Steps", minimum=6, maximum=25, step=1, value=8)
241
- scale = gr.Slider(label="Guidance Scale", minimum=0.0, maximum=5.0, step=0.1, value=3.5)
242
  seed = gr.Number(label="Seed (reproducibility)", value=1234, precision=0)
243
 
244
  generate_btn = gr.Button("Generate Base Image", variant="primary")
245
- generated_image = gr.Image(label="Generated Image", type="pil")
246
 
247
  with gr.Column():
248
- gr.Markdown("## 2) (Optional) Change Text in the Generated Image (Gemini)")
249
  new_text_input = gr.Textbox(
250
- label="New Text to Insert",
251
- placeholder="e.g. Hello"
252
  )
253
  modify_btn = gr.Button("Change Text in Image via Gemini", variant="secondary")
254
  output_img = gr.Image(label="Modified Image", type="pil")
 
78
  - file_name: ์›๋ณธ ์ด๋ฏธ์ง€(์˜ˆ: .png) ๊ฒฝ๋กœ
79
  - model: ์‚ฌ์šฉํ•  gemini ๋ชจ๋ธ ์ด๋ฆ„
80
  """
81
+ # (1) ํ™˜๊ฒฝ ๋ณ€์ˆ˜์—์„œ API ํ‚ค ๊ฐ€์ ธ์˜ค๊ธฐ (ํ•„์ˆ˜)
82
  api_key = os.getenv("GAPI_TOKEN", None)
83
  if not api_key:
84
  raise ValueError(
 
86
  "Google GenAI API๋ฅผ ์‚ฌ์šฉํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” GAPI_TOKEN์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค."
87
  )
88
 
89
+ # (2) Google Client ์ดˆ๊ธฐํ™”
90
  client = genai.Client(api_key=api_key)
91
 
92
+ # (3) ์ด๋ฏธ์ง€ ์—…๋กœ๋“œ
93
  files = [client.files.upload(file=file_name)]
94
 
95
+ # (4) gemini์— ์ „๋‹ฌํ•  Content ์ค€๋น„ (์ด๋ฏธ์ง€ + ํ”„๋กฌํ”„ํŠธ)
96
  contents = [
97
  types.Content(
98
  role="user",
 
106
  ),
107
  ]
108
 
109
+ # (5) ์ƒ์„ฑ/๋ณ€ํ™˜ ์„ค์ •
110
  generate_content_config = types.GenerateContentConfig(
111
  temperature=1,
112
  top_p=0.95,
 
119
  text_response = ""
120
  image_path = None
121
 
122
+ # ์ž„์‹œ ํŒŒ์ผ๋กœ ์ด๋ฏธ์ง€ ๋ฐ›์„ ์ค€๋น„
123
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
124
  temp_path = tmp.name
125
+ # ์‘๋‹ต ์ŠคํŠธ๋ฆผ์„ ๋ฐ›์œผ๋ฉด์„œ ์ด๋ฏธ์ง€/ํ…์ŠคํŠธ ๊ตฌ๋ถ„ ์ฒ˜๋ฆฌ
126
  for chunk in client.models.generate_content_stream(
127
  model=model,
128
  contents=contents,
 
132
  continue
133
  candidate = chunk.candidates[0].content.parts[0]
134
 
135
+ # inline_data๊ฐ€ ์žˆ์œผ๋ฉด ์ด๋ฏธ์ง€ ์‘๋‹ต
136
  if candidate.inline_data:
137
  save_binary_file(temp_path, candidate.inline_data.data)
138
  print(f"File of mime type {candidate.inline_data.mime_type} saved to: {temp_path}")
139
  image_path = temp_path
140
  break
141
  else:
142
+ # ์ด๋ฏธ์ง€ ์—†์ด ํ…์ŠคํŠธ๋งŒ ๋ฐ˜ํ™˜๋˜๋Š” ๊ฒฝ์šฐ
143
  text_response += chunk.text + "\n"
144
 
145
+ # ์—…๋กœ๋“œํ•œ File ๊ฐ์ฒด ์ œ๊ฑฐ
146
  del files
147
+
148
  return image_path, text_response
149
 
150
  #######################################
151
+ # 3. Gradio ํ•จ์ˆ˜
152
+ # (1) FLUX๋กœ ์ด๋ฏธ์ง€ ์ƒ์„ฑ -> (2) Google GenAI๋กœ ํ…์ŠคํŠธ ๊ต์ฒด
153
  #######################################
154
 
155
  def generate_initial_image(prompt, text, height, width, steps, scale, seed):
156
  """
157
+ FLUX ํŒŒ์ดํ”„๋ผ์ธ์„ ์‚ฌ์šฉํ•ด 'ํ…์ŠคํŠธ๊ฐ€ ํฌํ•จ๋œ ์ด๋ฏธ์ง€๋ฅผ' ๋จผ์ € ์ƒ์„ฑ.
158
+ - prompt ๋‚ด <text>๋ฅผ text๋กœ ์น˜ํ™˜
159
+ - <text>๊ฐ€ ์—†๋‹ค๋ฉด "with clear readable text that says '<text>'"๋ฅผ ์ž๋™ ๋ถ™์ž„
160
  """
161
  if "<text>" in prompt:
162
  combined_prompt = prompt.replace("<text>", text)
163
  else:
164
  combined_prompt = f"{prompt} with clear readable text that says '{text}'"
165
+
166
+ # ๋””๋ฒ„๊ทธ์šฉ: ์ตœ์ข… ๋“ค์–ด๊ฐ€๋Š” ํ”„๋กฌํ”„ํŠธ๋ฅผ ํ™•์ธ
167
+ print(f"[DEBUG] Final combined_prompt: {combined_prompt}")
168
 
169
  with torch.inference_mode(), torch.autocast("cuda", dtype=torch.bfloat16), timer("inference"):
170
  result = pipe(
 
181
 
182
  def change_text_in_image(original_image, new_text):
183
  """
184
+ Google GenAI์˜ gemini ๋ชจ๋ธ์„ ํ†ตํ•ด,
185
  ์—…๋กœ๋“œ๋œ ์ด๋ฏธ์ง€ ๋‚ด๋ถ€์˜ ๋ฌธ๊ตฌ๋ฅผ `new_text`๋กœ ๋ณ€๊ฒฝํ•ด์ฃผ๋Š” ํ•จ์ˆ˜.
186
  """
187
  try:
 
188
  with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
189
  original_path = tmp.name
190
  original_image.save(original_path)
 
196
  )
197
 
198
  if image_path:
199
+ # Gradio ๊ตฌ๋ฒ„์ „์—๋Š” decode_base64_to_image๊ฐ€ ์—†์œผ๋ฏ€๋กœ PIL๋กœ ์ฒ˜๋ฆฌ
200
  with open(image_path, "rb") as f:
201
  image_data = f.read()
202
  modified_img = Image.open(io.BytesIO(image_data))
203
  return modified_img, ""
204
  else:
205
+ # ์ด๋ฏธ์ง€๊ฐ€ ์—†์ด ํ…์ŠคํŠธ๋งŒ ๋ฐ˜ํ™˜๋œ ๊ฒฝ์šฐ
206
  return None, text_response
207
 
208
  except Exception as e:
209
  raise gr.Error(f"Error: {e}")
210
 
211
  #######################################
212
+ # 4. Gradio ์ธํ„ฐํŽ˜์ด์Šค ๊ตฌ์„ฑ
213
  #######################################
214
 
215
  with gr.Blocks(title="Flux + Google GenAI Text Replacement") as demo:
216
  gr.Markdown(
217
  """
218
+ # Flux ๊ธฐ๋ฐ˜ ์ด๋ฏธ์ง€ ์ƒ์„ฑ + Google GenAI๋ฅผ ํ†ตํ•œ ํ…์ŠคํŠธ ๋ณ€ํ™˜
219
+
220
+ **Usage**:
221
+ - You can include `<text>` in the prompt. For example:
222
+ `white cat with speech bubble says <text>`
223
+ - Then, type the actual text in "Text to Include in the Image" (ex: "Hello" or "์•ˆ๋…•").
224
+ - If `<text>` is not found in your prompt, the text will be automatically appended as:
225
+ `with clear readable text that says '<text>'`.
226
+ - Finally, you can optionally change the text again via Gemini.
 
 
227
 
228
  ---
229
  """
 
231
 
232
  with gr.Row():
233
  with gr.Column():
234
+ gr.Markdown("## 1) Step 1: FLUX๋กœ ํ…์ŠคํŠธ ํฌํ•จ ์ด๋ฏธ์ง€ ์ƒ์„ฑ")
235
  prompt_input = gr.Textbox(
236
  lines=3,
237
+ label="์ด๋ฏธ์ง€ ์žฅ๋ฉด/๋ฐฐ๊ฒฝ Prompt (use `<text>` placeholder if you like)",
238
+ placeholder="e.g. A white cat with speech bubble says <text>"
239
  )
240
  text_input = gr.Textbox(
241
  lines=1,
242
+ label="์ด๋ฏธ์ง€ ์•ˆ์— ๋“ค์–ด๊ฐˆ ํ…์ŠคํŠธ",
243
+ placeholder="e.g. Hello or ์•ˆ๋…•"
244
  )
245
+ with gr.Accordion("๊ณ ๊ธ‰ ์„ค์ • (ํ™•์žฅ)", open=False):
246
  height = gr.Slider(label="Height", minimum=256, maximum=1152, step=64, value=512)
247
  width = gr.Slider(label="Width", minimum=256, maximum=1152, step=64, value=512)
248
  steps = gr.Slider(label="Inference Steps", minimum=6, maximum=25, step=1, value=8)
249
+ scale = gr.Slider(label="Guidance Scale", minimum=0.0, maximum=10.0, step=0.5, value=3.5)
250
  seed = gr.Number(label="Seed (reproducibility)", value=1234, precision=0)
251
 
252
  generate_btn = gr.Button("Generate Base Image", variant="primary")
253
+ generated_image = gr.Image(label="Generated Image (with text)", type="pil")
254
 
255
  with gr.Column():
256
+ gr.Markdown("## 2) Step 2: ์ƒ์„ฑ๋œ ์ด๋ฏธ์ง€ ๋‚ด ํ…์ŠคํŠธ ์ˆ˜์ •")
257
  new_text_input = gr.Textbox(
258
+ label="์ƒˆ๋กœ ๋ฐ”๊ฟ€ ํ…์ŠคํŠธ",
259
+ placeholder="์˜ˆ) Hello world"
260
  )
261
  modify_btn = gr.Button("Change Text in Image via Gemini", variant="secondary")
262
  output_img = gr.Image(label="Modified Image", type="pil")