Update app.py
Browse files
app.py
CHANGED
@@ -18,15 +18,52 @@ load_dotenv()
|
|
18 |
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
|
19 |
logger = logging.getLogger(__name__)
|
20 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
def save_binary_file(file_name, data):
|
22 |
with open(file_name, "wb") as f:
|
23 |
f.write(data)
|
24 |
|
25 |
-
def preprocess_prompt(prompt, image1, image2, image3):
|
26 |
"""
|
27 |
ํ๋กฌํํธ๋ฅผ ์ฒ๋ฆฌํ๊ณ ๊ธฐ๋ฅ ๋ช
๋ น์ ํด์
|
28 |
"""
|
29 |
-
# ๊ธฐ์กด preprocess_prompt ํจ์ ์ฝ๋ ์ ์ง
|
30 |
# ์ด๋ฏธ์ง ์๋ ์ฐธ์กฐ ํ์ธ ๋ฐ ์ฒ๋ฆฌ
|
31 |
has_img1 = image1 is not None
|
32 |
has_img2 = image2 is not None
|
@@ -50,22 +87,10 @@ def preprocess_prompt(prompt, image1, image2, image3):
|
|
50 |
|
51 |
# ๊ธฐ๋ฅ ๋ช
๋ น ํด์
|
52 |
if "1. ์ด๋ฏธ์ง ๋ณ๊ฒฝ" in prompt:
|
53 |
-
|
54 |
-
desc_match = re.search(r'#1์ "(.*?)"์ผ๋ก ๋ฐ๊ฟ๋ผ', prompt)
|
55 |
-
if desc_match:
|
56 |
-
description = desc_match.group(1)
|
57 |
-
prompt = f"์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง๋ฅผ {description}์ผ๋ก ๋ณ๊ฒฝํด์ฃผ์ธ์. ์๋ณธ ์ด๋ฏธ์ง์ ์ฃผ์ ๋ด์ฉ์ ์ ์งํ๋ ์๋ก์ด ์คํ์ผ๊ณผ ๋ถ์๊ธฐ๋ก ์ฌํด์ํด์ฃผ์ธ์."
|
58 |
-
else:
|
59 |
-
prompt = "์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง๋ฅผ ์ฐฝ์์ ์ผ๋ก ๋ณํํด์ฃผ์ธ์. ๋ ์์ํ๊ณ ์์ ์ ์ธ ๋ฒ์ ์ผ๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
60 |
|
61 |
elif "2. ๊ธ์์ง์ฐ๊ธฐ" in prompt:
|
62 |
-
|
63 |
-
text_match = re.search(r'#1์์ "(.*?)"๋ฅผ ์ง์๋ผ', prompt)
|
64 |
-
if text_match:
|
65 |
-
text_to_remove = text_match.group(1)
|
66 |
-
prompt = f"์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง์์ '{text_to_remove}' ํ
์คํธ๋ฅผ ์ฐพ์ ์์ฐ์ค๋ฝ๊ฒ ์ ๊ฑฐํด์ฃผ์ธ์. ํ
์คํธ๊ฐ ์๋ ๋ถ๋ถ์ ๋ฐฐ๊ฒฝ๊ณผ ์กฐํ๋กญ๊ฒ ์ฑ์์ฃผ์ธ์."
|
67 |
-
else:
|
68 |
-
prompt = "์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง์์ ๋ชจ๋ ํ
์คํธ๋ฅผ ์ฐพ์ ์์ฐ์ค๋ฝ๊ฒ ์ ๊ฑฐํด์ฃผ์ธ์. ๊น๋ํ ์ด๋ฏธ์ง๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
69 |
|
70 |
elif "3. ์ผ๊ตด๋ฐ๊พธ๊ธฐ" in prompt:
|
71 |
prompt = "์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง์ ์ธ๋ฌผ ์ผ๊ตด์ ๋ ๋ฒ์งธ ์ด๋ฏธ์ง์ ์ผ๊ตด๋ก ์์ฐ์ค๋ฝ๊ฒ ๊ต์ฒดํด์ฃผ์ธ์. ์ผ๊ตด์ ํ์ ๊ณผ ํน์ง์ ๋ ๋ฒ์งธ ์ด๋ฏธ์ง๋ฅผ ๋ฐ๋ฅด๋, ๋๋จธ์ง ๋ถ๋ถ์ ์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง๋ฅผ ์ ์งํด์ฃผ์ธ์."
|
@@ -99,96 +124,112 @@ def preprocess_prompt(prompt, image1, image2, image3):
|
|
99 |
|
100 |
return prompt
|
101 |
|
102 |
-
def generate_with_images(prompt, images):
|
103 |
"""
|
104 |
๊ณต์ ๋ฌธ์์ ๊ธฐ๋ฐํ ์ฌ๋ฐ๋ฅธ API ํธ์ถ ๋ฐฉ์ ๊ตฌํ
|
|
|
105 |
"""
|
106 |
-
|
107 |
-
|
108 |
-
|
109 |
-
|
110 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
111 |
|
112 |
-
|
113 |
-
|
114 |
-
|
115 |
-
|
116 |
-
|
117 |
-
# ์ปจํ
์ธ ์ค๋น
|
118 |
-
contents = []
|
119 |
-
|
120 |
-
# ํ
์คํธ ํ๋กฌํํธ ์ถ๊ฐ
|
121 |
-
contents.append(prompt)
|
122 |
-
|
123 |
-
# ์ด๋ฏธ์ง ์ถ๊ฐ
|
124 |
-
for idx, img in enumerate(images, 1):
|
125 |
-
if img is not None:
|
126 |
-
contents.append(img)
|
127 |
-
logger.info(f"์ด๋ฏธ์ง #{idx} ์ถ๊ฐ๋จ")
|
128 |
-
|
129 |
-
# ์์ฑ ์ค์ - ๊ณต์ ๋ฌธ์์ ๋ฐ๋ผ responseModalities ์ค์
|
130 |
-
response = client.models.generate_content(
|
131 |
-
model="gemini-2.0-flash-exp-image-generation",
|
132 |
-
contents=contents,
|
133 |
-
config=types.GenerateContentConfig(
|
134 |
-
response_modalities=['Text', 'Image'],
|
135 |
-
temperature=1,
|
136 |
-
top_p=0.95,
|
137 |
-
top_k=40,
|
138 |
-
max_output_tokens=8192
|
139 |
-
)
|
140 |
-
)
|
141 |
-
|
142 |
-
# ์์ ํ์ผ ์์ฑ
|
143 |
-
with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
|
144 |
-
temp_path = tmp.name
|
145 |
|
146 |
-
|
147 |
-
|
148 |
|
149 |
-
#
|
150 |
-
|
151 |
-
|
152 |
-
|
153 |
-
|
154 |
-
|
155 |
-
|
156 |
-
|
157 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
158 |
|
159 |
-
|
160 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
161 |
|
162 |
-
|
163 |
-
result_img = Image.open(temp_path)
|
164 |
-
if result_img.mode == "RGBA":
|
165 |
-
result_img = result_img.convert("RGB")
|
166 |
|
167 |
-
|
168 |
-
|
169 |
-
|
170 |
-
|
171 |
-
|
|
|
|
|
172 |
|
173 |
-
def process_images_with_prompt(image1, image2, image3, prompt):
|
174 |
"""
|
175 |
3๊ฐ์ ์ด๋ฏธ์ง์ ํ๋กฌํํธ๋ฅผ ์ฒ๋ฆฌํ๋ ํจ์
|
176 |
"""
|
177 |
try:
|
178 |
# ์ด๋ฏธ์ง ๊ฐ์ ํ์ธ
|
179 |
-
images = [image1, image2, image3]
|
180 |
-
valid_images = [img for img in images if img is not None]
|
181 |
-
|
182 |
-
if not valid_images:
|
183 |
-
return None, "์ ์ด๋ ํ๋์ ์ด๋ฏธ์ง๋ฅผ ์
๋ก๋ํด์ฃผ์ธ์."
|
184 |
|
185 |
# ํ๋กฌํํธ ์ฒ๋ฆฌ
|
186 |
if not prompt or not prompt.strip():
|
187 |
# ํ๋กฌํํธ๊ฐ ์์ผ๋ฉด ์
๋ก๋๋ ์ด๋ฏธ์ง ์์ ๋ฐ๋ผ ์๋ ํฉ์ฑ ํ๋กฌํํธ ์์ฑ
|
188 |
-
if len(
|
|
|
|
|
|
|
189 |
prompt = "์ด ์ด๋ฏธ์ง๋ฅผ ์ฐฝ์์ ์ผ๋ก ๋ณํํด์ฃผ์ธ์. ๋ ์์ํ๊ณ ์์ ์ ์ธ ๋ฒ์ ์ผ๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
190 |
logger.info("๋จ์ผ ์ด๋ฏธ์ง ํ๋กฌํํธ ์๋ ์์ฑ")
|
191 |
-
elif len(
|
192 |
prompt = "์ด ๋ ์ด๋ฏธ์ง๋ฅผ ์์ฐ์ค๋ฝ๊ฒ ํฉ์ฑํด์ฃผ์ธ์. ๋ ์ด๋ฏธ์ง์ ์์๋ฅผ ์กฐํ๋กญ๊ฒ ํตํฉํ์ฌ ํ๋์ ์ด๋ฏธ์ง๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
193 |
logger.info("๋ ์ด๋ฏธ์ง ํฉ์ฑ ํ๋กฌํํธ ์๋ ์์ฑ")
|
194 |
else:
|
@@ -198,18 +239,18 @@ def process_images_with_prompt(image1, image2, image3, prompt):
|
|
198 |
# ํ๋กฌํํธ ์ ์ฒ๋ฆฌ ๋ฐ ๊ธฐ๋ฅ ๋ช
๋ น ํด์
|
199 |
prompt = preprocess_prompt(prompt, image1, image2, image3)
|
200 |
|
201 |
-
#
|
202 |
-
return generate_with_images(prompt,
|
203 |
|
204 |
except Exception as e:
|
205 |
logger.exception("์ด๋ฏธ์ง ์ฒ๋ฆฌ ์ค ์ค๋ฅ ๋ฐ์:")
|
206 |
return None, f"์ค๋ฅ ๋ฐ์: {str(e)}"
|
207 |
|
208 |
-
# ๊ธฐ๋ฅ ์ ํ ์ฝ๋ฐฑ (
|
209 |
-
def update_prompt_from_function(function_choice
|
210 |
function_templates = {
|
211 |
-
"1. ์ด๋ฏธ์ง ๋ณ๊ฒฝ":
|
212 |
-
"2. ๊ธ์์ง์ฐ๊ธฐ":
|
213 |
"3. ์ผ๊ตด๋ฐ๊พธ๊ธฐ": "#1์ ์ธ๋ฌผ์ #2์ ์ผ๊ตด๋ก ๋ฐ๊ฟ๋ผ",
|
214 |
"4. ์ท๋ฐ๊พธ๊ธฐ": "#1์ ์ธ๋ฌผ์ #2 ๋๋ #3์ ์ท์ผ๋ก ๋ณ๊ฒฝํ๋ผ",
|
215 |
"5. ๋ฐฐ๊ฒฝ๋ฐ๊พธ๊ธฐ": "#1์ ์ด๋ฏธ์ง์ #2์ ๋ฐฐ๊ฒฝ์ผ๋ก ์์ฐ์ค๋ฝ๊ฒ ๋ฐ๊ฟ๋ผ",
|
@@ -219,119 +260,107 @@ def update_prompt_from_function(function_choice, custom_text=""):
|
|
219 |
|
220 |
return function_templates.get(function_choice, "")
|
221 |
|
222 |
-
# Gradio ์ธํฐํ์ด์ค (
|
223 |
-
|
224 |
-
|
225 |
-
|
226 |
-
<div style="text-align: center; margin-bottom: 1rem;">
|
227 |
-
<h1>๊ฐ๋จํ ์ด๋ฏธ์ง ์์ฑ๊ธฐ</h1>
|
228 |
-
<p>์ด๋ฏธ์ง๋ฅผ ์
๋ก๋ํ๊ณ ๋ฐ๋ก ์คํํ๋ฉด ์๋์ผ๋ก ํฉ์ฑํฉ๋๋ค. ๋๋ ์๋ ๊ธฐ๋ฅ์ ์ ํํ ์ ์์ต๋๋ค.</p>
|
229 |
-
</div>
|
230 |
-
"""
|
231 |
-
)
|
232 |
|
233 |
-
with gr.
|
234 |
-
|
235 |
-
|
236 |
-
|
237 |
-
|
238 |
-
|
239 |
-
|
240 |
-
|
241 |
-
|
242 |
-
|
243 |
-
|
244 |
-
|
245 |
-
|
246 |
-
|
247 |
-
|
248 |
-
|
249 |
-
|
250 |
-
|
251 |
-
|
252 |
-
|
253 |
-
|
254 |
-
|
255 |
-
|
256 |
-
|
257 |
-
|
258 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
259 |
)
|
|
|
|
|
|
|
260 |
|
261 |
-
|
262 |
-
|
263 |
-
|
264 |
-
|
265 |
-
|
266 |
-
|
267 |
-
label="ํ๋กฌํํธ
|
268 |
-
)
|
269 |
-
|
270 |
-
# ์์ฑ ๋ฒํผ
|
271 |
-
submit_btn = gr.Button("์ด๋ฏธ์ง ์์ฑ", variant="primary")
|
272 |
|
273 |
-
|
274 |
-
|
275 |
-
|
276 |
-
|
277 |
-
|
278 |
-
|
279 |
-
prompt_display = gr.Textbox(label="์ฌ์ฉ๋ ํ๋กฌํํธ", visible=True)
|
280 |
-
|
281 |
-
# ๊ธฐ๋ฅ ์ ์ฉ ๋ฒํผ ์ด๋ฒคํธ
|
282 |
-
apply_function_btn.click(
|
283 |
-
fn=update_prompt_from_function,
|
284 |
-
inputs=[function_dropdown, custom_text_input],
|
285 |
-
outputs=[prompt_input]
|
286 |
-
)
|
287 |
-
|
288 |
-
# ์ด๋ฏธ์ง ์์ฑ ๋ฒํผ ํด๋ฆญ ์ด๋ฒคํธ
|
289 |
-
def process_and_show_prompt(image1, image2, image3, prompt):
|
290 |
-
# ์ด๋ฏธ์ง ๊ฐ์ ํ์ธ
|
291 |
-
images = [image1, image2, image3]
|
292 |
-
valid_images = [img for img in images if img is not None]
|
293 |
|
294 |
-
|
295 |
-
|
296 |
-
|
297 |
-
|
298 |
-
|
299 |
-
|
300 |
-
|
301 |
-
|
302 |
-
|
303 |
-
|
304 |
-
|
305 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
306 |
|
307 |
-
|
308 |
-
|
|
|
|
|
|
|
309 |
|
310 |
-
|
311 |
-
|
312 |
-
|
313 |
-
return None, f"์ค๋ฅ ๋ฐ์: {str(e)}", prompt
|
314 |
|
315 |
-
|
316 |
-
fn=process_and_show_prompt,
|
317 |
-
inputs=[image1_input, image2_input, image3_input, prompt_input],
|
318 |
-
outputs=[output_image, output_text, prompt_display],
|
319 |
-
)
|
320 |
-
|
321 |
-
gr.Markdown(
|
322 |
-
"""
|
323 |
-
### ์ฌ์ฉ ๋ฐฉ๋ฒ:
|
324 |
-
|
325 |
-
1. **์๋ ํฉ์ฑ**: ์ด๋ฏธ์ง๋ง ์
๋ก๋ํ๊ณ ํ๋กฌํํธ๋ฅผ ๋น์๋๋ฉด ์๋์ผ๋ก ํฉ์ฑ๋ฉ๋๋ค
|
326 |
-
2. **๊ธฐ๋ฅ ์ฌ์ฉ**: ๋๋กญ๋ค์ด์์ ์ํ๋ ๊ธฐ๋ฅ์ ์ ํํ๊ณ '๊ธฐ๋ฅ ์ ์ฉ' ๋ฒํผ์ ํด๋ฆญํ์ธ์
|
327 |
-
3. **์ปค์คํ
ํ
์คํธ**: ์ด๋ฏธ์ง ๋ณ๊ฒฝ์ด๋ ๊ธ์์ง์ฐ๊ธฐ ๊ธฐ๋ฅ์ ์ ํํ ๋ ์ถ๊ฐ ์ค๋ช
์ ์
๋ ฅํ ์ ์์ต๋๋ค
|
328 |
-
4. **์ด๋ฏธ์ง ์ฐธ์กฐ**: #1, #2, #3์ผ๋ก ๊ฐ ์ด๋ฏธ์ง๋ฅผ ์ฐธ์กฐํ ์ ์์ต๋๋ค
|
329 |
-
5. **์ผ๋ถ ์ด๋ฏธ์ง๋ง**: ํ์ํ ์ด๋ฏธ์ง๋ง ์
๋ก๋ํด๋ ๊ธฐ๋ฅ ์คํ์ด ๊ฐ๋ฅํฉ๋๋ค
|
330 |
-
|
331 |
-
> **ํ**: ๊ธฐ๋ฅ์ ์ ํํ ํ์๋ ํ๋กฌํํธ๋ฅผ ์ง์ ์์ ํ ์ ์์ต๋๋ค
|
332 |
-
"""
|
333 |
-
)
|
334 |
|
335 |
# ์ ํ๋ฆฌ์ผ์ด์
์คํ
|
336 |
if __name__ == "__main__":
|
|
|
337 |
demo.launch(share=True)
|
|
|
18 |
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
|
19 |
logger = logging.getLogger(__name__)
|
20 |
|
21 |
+
# LLM ๋ชจ๋ธ ์ด๊ธฐํ (๋ฒ์ญ ๊ธฐ๋ฅ์ฉ)
|
22 |
+
def initialize_gemini():
|
23 |
+
api_key = os.environ.get("GEMINI_API_KEY")
|
24 |
+
if not api_key:
|
25 |
+
logger.error("GEMINI_API_KEY๊ฐ ์ค์ ๋์ง ์์์ต๋๋ค.")
|
26 |
+
return None
|
27 |
+
|
28 |
+
genai.configure(api_key=api_key)
|
29 |
+
return genai.GenerativeModel(
|
30 |
+
model_name="gemini-2.0-flash",
|
31 |
+
generation_config={
|
32 |
+
"temperature": 0.2, # ๋ฎ์ temperature๋ก ์ ํํ ๋ฒ์ญ ์ ๋
|
33 |
+
"max_output_tokens": 1024,
|
34 |
+
"top_p": 0.9,
|
35 |
+
}
|
36 |
+
)
|
37 |
+
|
38 |
+
# LLM ๋ชจ๋ธ๋ก ํ๊ตญ์ด ํ๋กฌํํธ๋ฅผ ์์ด๋ก ๋ฒ์ญ
|
39 |
+
def translate_to_english(korean_prompt, model):
|
40 |
+
if not korean_prompt or not model:
|
41 |
+
return korean_prompt
|
42 |
+
|
43 |
+
try:
|
44 |
+
translation_prompt = f"""
|
45 |
+
๋ค์ ํ๊ตญ์ด ํ
์คํธ๋ฅผ ์์ด๋ก ์ ํํ๊ฒ ๋ฒ์ญํด์ฃผ์ธ์. ๋งฅ๋ฝ๊ณผ ์๋ฏธ๋ฅผ ์ต๋ํ ๋ณด์กดํ์ธ์.
|
46 |
+
๋ฒ์ญํ ํ
์คํธ: {korean_prompt}
|
47 |
+
|
48 |
+
๋ฒ์ญ๋ง ์ ๊ณตํ๊ณ ๋ค๋ฅธ ์ค๋ช
์ ํ์ง ๋ง์ธ์.
|
49 |
+
"""
|
50 |
+
|
51 |
+
response = model.generate_content(translation_prompt)
|
52 |
+
english_prompt = response.text.strip()
|
53 |
+
logger.info(f"๋ฒ์ญ ๊ฒฐ๊ณผ: {english_prompt}")
|
54 |
+
return english_prompt
|
55 |
+
except Exception as e:
|
56 |
+
logger.exception(f"๋ฒ์ญ ์ค ์ค๋ฅ ๋ฐ์: {str(e)}")
|
57 |
+
return korean_prompt # ์ค๋ฅ ๋ฐ์ ์ ์๋ณธ ํ๊ตญ์ด ํ๋กฌํํธ ๋ฐํ
|
58 |
+
|
59 |
def save_binary_file(file_name, data):
|
60 |
with open(file_name, "wb") as f:
|
61 |
f.write(data)
|
62 |
|
63 |
+
def preprocess_prompt(prompt, image1=None, image2=None, image3=None):
|
64 |
"""
|
65 |
ํ๋กฌํํธ๋ฅผ ์ฒ๋ฆฌํ๊ณ ๊ธฐ๋ฅ ๋ช
๋ น์ ํด์
|
66 |
"""
|
|
|
67 |
# ์ด๋ฏธ์ง ์๋ ์ฐธ์กฐ ํ์ธ ๋ฐ ์ฒ๋ฆฌ
|
68 |
has_img1 = image1 is not None
|
69 |
has_img2 = image2 is not None
|
|
|
87 |
|
88 |
# ๊ธฐ๋ฅ ๋ช
๋ น ํด์
|
89 |
if "1. ์ด๋ฏธ์ง ๋ณ๊ฒฝ" in prompt:
|
90 |
+
prompt = "์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง๋ฅผ ์ฐฝ์์ ์ผ๋ก ๋ณํํด์ฃผ์ธ์. ๋ ์์ํ๊ณ ์์ ์ ์ธ ๋ฒ์ ์ผ๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
|
92 |
elif "2. ๊ธ์์ง์ฐ๊ธฐ" in prompt:
|
93 |
+
prompt = "์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง์์ ๋ชจ๋ ํ
์คํธ๋ฅผ ์ฐพ์ ์์ฐ์ค๋ฝ๊ฒ ์ ๊ฑฐํด์ฃผ์ธ์. ๊น๋ํ ์ด๋ฏธ์ง๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
|
|
|
|
|
|
|
|
|
|
|
|
94 |
|
95 |
elif "3. ์ผ๊ตด๋ฐ๊พธ๊ธฐ" in prompt:
|
96 |
prompt = "์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง์ ์ธ๋ฌผ ์ผ๊ตด์ ๋ ๋ฒ์งธ ์ด๋ฏธ์ง์ ์ผ๊ตด๋ก ์์ฐ์ค๋ฝ๊ฒ ๊ต์ฒดํด์ฃผ์ธ์. ์ผ๊ตด์ ํ์ ๊ณผ ํน์ง์ ๋ ๋ฒ์งธ ์ด๋ฏธ์ง๋ฅผ ๋ฐ๋ฅด๋, ๋๋จธ์ง ๋ถ๋ถ์ ์ฒซ ๋ฒ์งธ ์ด๋ฏธ์ง๋ฅผ ์ ์งํด์ฃผ์ธ์."
|
|
|
124 |
|
125 |
return prompt
|
126 |
|
127 |
+
def generate_with_images(prompt, images, gemini_model):
|
128 |
"""
|
129 |
๊ณต์ ๋ฌธ์์ ๊ธฐ๋ฐํ ์ฌ๋ฐ๋ฅธ API ํธ์ถ ๋ฐฉ์ ๊ตฌํ
|
130 |
+
์ฌ์๋ ๋ก์ง ์ถ๊ฐ
|
131 |
"""
|
132 |
+
max_retries = 2
|
133 |
+
retries = 0
|
134 |
+
|
135 |
+
while retries <= max_retries:
|
136 |
+
try:
|
137 |
+
# API ํค ํ์ธ
|
138 |
+
api_key = os.environ.get("GEMINI_API_KEY")
|
139 |
+
if not api_key:
|
140 |
+
return None, "API ํค๊ฐ ์ค์ ๋์ง ์์์ต๋๋ค. ํ๊ฒฝ๋ณ์๋ฅผ ํ์ธํด์ฃผ์ธ์."
|
141 |
+
|
142 |
+
# Gemini ํด๋ผ์ด์ธํธ ์ด๊ธฐํ
|
143 |
+
client = genai.Client(api_key=api_key)
|
144 |
|
145 |
+
# ํ๋กฌํํธ ๋ฒ์ญ
|
146 |
+
english_prompt = translate_to_english(prompt, gemini_model) if gemini_model else prompt
|
147 |
+
logger.info(f"์๋ณธ ํ๋กฌํํธ: {prompt}")
|
148 |
+
logger.info(f"๋ฒ์ญ๋ ํ๋กฌํํธ: {english_prompt}")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
149 |
|
150 |
+
# ์ปจํ
์ธ ์ค๋น
|
151 |
+
contents = []
|
152 |
|
153 |
+
# ํ
์คํธ ํ๋กฌํํธ ์ถ๊ฐ
|
154 |
+
contents.append(english_prompt)
|
155 |
+
|
156 |
+
# ์ด๋ฏธ์ง ์ถ๊ฐ
|
157 |
+
for idx, img in enumerate(images, 1):
|
158 |
+
if img is not None:
|
159 |
+
contents.append(img)
|
160 |
+
logger.info(f"์ด๋ฏธ์ง #{idx} ์ถ๊ฐ๋จ")
|
161 |
+
|
162 |
+
# ์์ฑ ์ค์ - ๊ณต์ ๋ฌธ์์ ๋ฐ๋ผ responseModalities ์ค์
|
163 |
+
response = client.models.generate_content(
|
164 |
+
model="gemini-2.0-flash-exp-image-generation",
|
165 |
+
contents=contents,
|
166 |
+
config=types.GenerateContentConfig(
|
167 |
+
response_modalities=['Text', 'Image'],
|
168 |
+
temperature=1,
|
169 |
+
top_p=0.95,
|
170 |
+
top_k=40,
|
171 |
+
max_output_tokens=8192
|
172 |
+
)
|
173 |
+
)
|
174 |
|
175 |
+
# ์์ ํ์ผ ์์ฑ
|
176 |
+
with tempfile.NamedTemporaryFile(suffix=".png", delete=False) as tmp:
|
177 |
+
temp_path = tmp.name
|
178 |
+
|
179 |
+
result_text = ""
|
180 |
+
image_found = False
|
181 |
+
|
182 |
+
# ์๋ต ์ฒ๋ฆฌ
|
183 |
+
for part in response.candidates[0].content.parts:
|
184 |
+
if hasattr(part, 'text') and part.text:
|
185 |
+
result_text += part.text
|
186 |
+
logger.info(f"์๋ต ํ
์คํธ: {part.text}")
|
187 |
+
elif hasattr(part, 'inline_data') and part.inline_data:
|
188 |
+
save_binary_file(temp_path, part.inline_data.data)
|
189 |
+
image_found = True
|
190 |
+
logger.info("์๋ต์์ ์ด๋ฏธ์ง ์ถ์ถ ์ฑ๊ณต")
|
191 |
+
|
192 |
+
if not image_found:
|
193 |
+
if retries < max_retries:
|
194 |
+
retries += 1
|
195 |
+
logger.warning(f"์ด๋ฏธ์ง ์์ฑ ์คํจ, ์ฌ์๋ {retries}/{max_retries}")
|
196 |
+
continue
|
197 |
+
else:
|
198 |
+
return None, f"API์์ ์ด๋ฏธ์ง๋ฅผ ์์ฑํ์ง ๋ชปํ์ต๋๋ค. ์๋ต ํ
์คํธ: {result_text}"
|
199 |
+
|
200 |
+
# ๊ฒฐ๊ณผ ์ด๋ฏธ์ง ๋ฐํ
|
201 |
+
result_img = Image.open(temp_path)
|
202 |
+
if result_img.mode == "RGBA":
|
203 |
+
result_img = result_img.convert("RGB")
|
204 |
|
205 |
+
return result_img, f"์ด๋ฏธ์ง๊ฐ ์ฑ๊ณต์ ์ผ๋ก ์์ฑ๋์์ต๋๋ค. {result_text}"
|
|
|
|
|
|
|
206 |
|
207 |
+
except Exception as e:
|
208 |
+
if retries < max_retries:
|
209 |
+
retries += 1
|
210 |
+
logger.warning(f"์ด๋ฏธ์ง ์์ฑ ์ค ์ค๋ฅ ๋ฐ์, ์ฌ์๋ {retries}/{max_retries}: {str(e)}")
|
211 |
+
else:
|
212 |
+
logger.exception("์ด๋ฏธ์ง ์์ฑ ์ค ์ค๋ฅ ๋ฐ์:")
|
213 |
+
return None, f"์ค๋ฅ ๋ฐ์: {str(e)}"
|
214 |
|
215 |
+
def process_images_with_prompt(image1, image2, image3, prompt, gemini_model):
|
216 |
"""
|
217 |
3๊ฐ์ ์ด๋ฏธ์ง์ ํ๋กฌํํธ๋ฅผ ์ฒ๋ฆฌํ๋ ํจ์
|
218 |
"""
|
219 |
try:
|
220 |
# ์ด๋ฏธ์ง ๊ฐ์ ํ์ธ
|
221 |
+
images = [img for img in [image1, image2, image3] if img is not None]
|
|
|
|
|
|
|
|
|
222 |
|
223 |
# ํ๋กฌํํธ ์ฒ๋ฆฌ
|
224 |
if not prompt or not prompt.strip():
|
225 |
# ํ๋กฌํํธ๊ฐ ์์ผ๋ฉด ์
๋ก๋๋ ์ด๋ฏธ์ง ์์ ๋ฐ๋ผ ์๋ ํฉ์ฑ ํ๋กฌํํธ ์์ฑ
|
226 |
+
if len(images) == 0:
|
227 |
+
prompt = "์๋ฆ๋ค์ด ์์ฐ ํ๊ฒฝ ์ด๋ฏธ์ง๋ฅผ ์์ฑํด์ฃผ์ธ์. ์ฐ, ํธ์, ํ๋์ด ํฌํจ๋ ํํ๋ก์ด ์ฅ๋ฉด์ด๋ฉด ์ข๊ฒ ์ต๋๋ค."
|
228 |
+
logger.info("์ด๋ฏธ์ง ์์, ๊ธฐ๋ณธ ์์ฑ ํ๋กฌํํธ ์ฌ์ฉ")
|
229 |
+
elif len(images) == 1:
|
230 |
prompt = "์ด ์ด๋ฏธ์ง๋ฅผ ์ฐฝ์์ ์ผ๋ก ๋ณํํด์ฃผ์ธ์. ๋ ์์ํ๊ณ ์์ ์ ์ธ ๋ฒ์ ์ผ๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
231 |
logger.info("๋จ์ผ ์ด๋ฏธ์ง ํ๋กฌํํธ ์๋ ์์ฑ")
|
232 |
+
elif len(images) == 2:
|
233 |
prompt = "์ด ๋ ์ด๋ฏธ์ง๋ฅผ ์์ฐ์ค๋ฝ๊ฒ ํฉ์ฑํด์ฃผ์ธ์. ๋ ์ด๋ฏธ์ง์ ์์๋ฅผ ์กฐํ๋กญ๊ฒ ํตํฉํ์ฌ ํ๋์ ์ด๋ฏธ์ง๋ก ๋ง๋ค์ด์ฃผ์ธ์."
|
234 |
logger.info("๋ ์ด๋ฏธ์ง ํฉ์ฑ ํ๋กฌํํธ ์๋ ์์ฑ")
|
235 |
else:
|
|
|
239 |
# ํ๋กฌํํธ ์ ์ฒ๋ฆฌ ๋ฐ ๊ธฐ๋ฅ ๋ช
๋ น ํด์
|
240 |
prompt = preprocess_prompt(prompt, image1, image2, image3)
|
241 |
|
242 |
+
# ์ด๋ฏธ์ง ์์ฑ API ํธ์ถ
|
243 |
+
return generate_with_images(prompt, images, gemini_model)
|
244 |
|
245 |
except Exception as e:
|
246 |
logger.exception("์ด๋ฏธ์ง ์ฒ๋ฆฌ ์ค ์ค๋ฅ ๋ฐ์:")
|
247 |
return None, f"์ค๋ฅ ๋ฐ์: {str(e)}"
|
248 |
|
249 |
+
# ๊ธฐ๋ฅ ์ ํ ์ฝ๋ฐฑ (์์ ๋ ๋ฒ์ )
|
250 |
+
def update_prompt_from_function(function_choice):
|
251 |
function_templates = {
|
252 |
+
"1. ์ด๋ฏธ์ง ๋ณ๊ฒฝ": "#1์ ์ฐฝ์์ ์ผ๋ก ๋ณํํด์ฃผ์ธ์",
|
253 |
+
"2. ๊ธ์์ง์ฐ๊ธฐ": "#1์์ ํ
์คํธ๋ฅผ ์ง์์ฃผ์ธ์",
|
254 |
"3. ์ผ๊ตด๋ฐ๊พธ๊ธฐ": "#1์ ์ธ๋ฌผ์ #2์ ์ผ๊ตด๋ก ๋ฐ๊ฟ๋ผ",
|
255 |
"4. ์ท๋ฐ๊พธ๊ธฐ": "#1์ ์ธ๋ฌผ์ #2 ๋๋ #3์ ์ท์ผ๋ก ๋ณ๊ฒฝํ๋ผ",
|
256 |
"5. ๋ฐฐ๊ฒฝ๋ฐ๊พธ๊ธฐ": "#1์ ์ด๋ฏธ์ง์ #2์ ๋ฐฐ๊ฒฝ์ผ๋ก ์์ฐ์ค๋ฝ๊ฒ ๋ฐ๊ฟ๋ผ",
|
|
|
260 |
|
261 |
return function_templates.get(function_choice, "")
|
262 |
|
263 |
+
# Gradio ์ธํฐํ์ด์ค (์์ ๋ ๋ฒ์ )
|
264 |
+
def create_interface():
|
265 |
+
# Gemini ๋ชจ๋ธ ์ด๊ธฐํ
|
266 |
+
gemini_model = initialize_gemini()
|
|
|
|
|
|
|
|
|
|
|
|
|
267 |
|
268 |
+
with gr.Blocks() as demo:
|
269 |
+
gr.HTML(
|
270 |
+
"""
|
271 |
+
<div style="text-align: center; margin-bottom: 1rem;">
|
272 |
+
<h1>๊ฐ๋จํ ์ด๋ฏธ์ง ์์ฑ๊ธฐ</h1>
|
273 |
+
<p>์ด๋ฏธ์ง๋ฅผ ์
๋ก๋ํ๊ฑฐ๋ ํ๋กฌํํธ๋ง ์
๋ ฅํ์ฌ ์ด๋ฏธ์ง๋ฅผ ์์ฑํ ์ ์์ต๋๋ค.</p>
|
274 |
+
</div>
|
275 |
+
"""
|
276 |
+
)
|
277 |
+
|
278 |
+
with gr.Row():
|
279 |
+
with gr.Column():
|
280 |
+
# 3๊ฐ์ ์ด๋ฏธ์ง ์
๋ ฅ
|
281 |
+
with gr.Row():
|
282 |
+
image1_input = gr.Image(type="pil", label="#1 (์ ํ์ฌํญ)", image_mode="RGB")
|
283 |
+
image2_input = gr.Image(type="pil", label="#2 (์ ํ์ฌํญ)", image_mode="RGB")
|
284 |
+
image3_input = gr.Image(type="pil", label="#3 (์ ํ์ฌํญ)", image_mode="RGB")
|
285 |
+
|
286 |
+
# ๊ธฐ๋ฅ ์ ํ ๋๋กญ๋ค์ด
|
287 |
+
with gr.Row():
|
288 |
+
function_dropdown = gr.Dropdown(
|
289 |
+
choices=[
|
290 |
+
"1. ์ด๋ฏธ์ง ๋ณ๊ฒฝ",
|
291 |
+
"2. ๊ธ์์ง์ฐ๊ธฐ",
|
292 |
+
"3. ์ผ๊ตด๋ฐ๊พธ๊ธฐ",
|
293 |
+
"4. ์ท๋ฐ๊พธ๊ธฐ",
|
294 |
+
"5. ๋ฐฐ๊ฒฝ๋ฐ๊พธ๊ธฐ",
|
295 |
+
"6. ์ด๋ฏธ์ง ํฉ์ฑ(์ํํฌํจ)",
|
296 |
+
"7. ์ด๋ฏธ์ง ํฉ์ฑ(์คํ์ผ์ ์ฉ)"
|
297 |
+
],
|
298 |
+
label="๊ธฐ๋ฅ ์ ํ (์ ํํ๋ฉด ์๋์ผ๋ก ํ๋กฌํํธ๊ฐ ๋ณ๊ฒฝ๋ฉ๋๋ค)",
|
299 |
+
value=None
|
300 |
+
)
|
301 |
+
|
302 |
+
# ํ๋กฌํํธ ์
๋ ฅ
|
303 |
+
prompt_input = gr.Textbox(
|
304 |
+
lines=3,
|
305 |
+
placeholder="ํ๋กฌํํธ๋ฅผ ์
๋ ฅํ๊ฑฐ๋ ๋น์๋๋ฉด ์๋ ์์ฑ๋ฉ๋๋ค. ํ๊ตญ์ด๋ก ์
๋ ฅํ์ธ์.",
|
306 |
+
label="ํ๋กฌํํธ"
|
307 |
)
|
308 |
+
|
309 |
+
# ์์ฑ ๋ฒํผ
|
310 |
+
submit_btn = gr.Button("์ด๋ฏธ์ง ์์ฑ", variant="primary")
|
311 |
|
312 |
+
with gr.Column():
|
313 |
+
# ๊ฒฐ๊ณผ ์ถ๋ ฅ
|
314 |
+
output_image = gr.Image(label="์์ฑ๋ ์ด๋ฏธ์ง")
|
315 |
+
output_text = gr.Textbox(label="์ํ ๋ฉ์์ง")
|
316 |
+
|
317 |
+
# ์ฌ์ฉ๋ ํ๋กฌํํธ ํ์
|
318 |
+
prompt_display = gr.Textbox(label="์ฌ์ฉ๋ ํ๋กฌํํธ", visible=True)
|
|
|
|
|
|
|
|
|
319 |
|
320 |
+
# ๊ธฐ๋ฅ ์ ํ ์ด๋ฒคํธ (์ ํ๋ง์ผ๋ก ํ๋กฌํํธ ๋ณ๊ฒฝ)
|
321 |
+
function_dropdown.change(
|
322 |
+
fn=update_prompt_from_function,
|
323 |
+
inputs=[function_dropdown],
|
324 |
+
outputs=[prompt_input]
|
325 |
+
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
326 |
|
327 |
+
# ์ด๋ฏธ์ง ์์ฑ ๋ฒํผ ํด๋ฆญ ์ด๋ฒคํธ
|
328 |
+
def process_and_show_prompt(image1, image2, image3, prompt):
|
329 |
+
try:
|
330 |
+
# ์ด๋ฏธ์ง ์์ฑ ํจ์ ํธ์ถ
|
331 |
+
result_img, status = process_images_with_prompt(image1, image2, image3, prompt, gemini_model)
|
332 |
+
|
333 |
+
# ํ๋กฌํํธ ์ ์ฒ๋ฆฌ
|
334 |
+
processed_prompt = preprocess_prompt(prompt, image1, image2, image3)
|
335 |
+
|
336 |
+
return result_img, status, processed_prompt
|
337 |
+
except Exception as e:
|
338 |
+
logger.exception("์ฒ๋ฆฌ ์ค ์ค๋ฅ ๋ฐ์:")
|
339 |
+
return None, f"์ค๋ฅ ๋ฐ์: {str(e)}", prompt
|
340 |
+
|
341 |
+
submit_btn.click(
|
342 |
+
fn=process_and_show_prompt,
|
343 |
+
inputs=[image1_input, image2_input, image3_input, prompt_input],
|
344 |
+
outputs=[output_image, output_text, prompt_display],
|
345 |
+
)
|
346 |
+
|
347 |
+
gr.Markdown(
|
348 |
+
"""
|
349 |
+
### ์ฌ์ฉ ๋ฐฉ๋ฒ:
|
350 |
|
351 |
+
1. **์ด๋ฏธ์ง ์์ด ์์ฑ**: ํ๋กฌํํธ๋ง ์
๋ ฅํ์ฌ ์ด๋ฏธ์ง๋ฅผ ์์ฑํ ์ ์์ต๋๋ค
|
352 |
+
2. **์๋ ํฉ์ฑ**: ์ด๋ฏธ์ง๋ง ์
๋ก๋ํ๊ณ ํ๋กฌํํธ๋ฅผ ๋น์๋๋ฉด ์๋์ผ๋ก ํฉ์ฑ๋ฉ๋๋ค
|
353 |
+
3. **๊ธฐ๋ฅ ์ฌ์ฉ**: ๋๋กญ๋ค์ด์์ ์ํ๋ ๊ธฐ๋ฅ์ ์ ํํ๋ฉด ์๋์ผ๋ก ํ๋กฌํํธ๊ฐ ๋ณ๊ฒฝ๋ฉ๋๋ค
|
354 |
+
4. **์ด๋ฏธ์ง ์ฐธ์กฐ**: #1, #2, #3์ผ๋ก ๊ฐ ์ด๋ฏธ์ง๋ฅผ ์ฐธ์กฐํ ์ ์์ต๋๋ค
|
355 |
+
5. **์ผ๋ถ ์ด๋ฏธ์ง๋ง**: ํ์ํ ์ด๋ฏธ์ง๋ง ์
๋ก๋ํด๋ ๊ธฐ๋ฅ ์คํ์ด ๊ฐ๋ฅํฉ๋๋ค
|
356 |
|
357 |
+
> **ํ**: ํ๊ตญ์ด๋ก ์
๋ ฅํ ํ๋กฌํํธ๋ ์๋์ผ๋ก ์์ด๋ก ๋ฒ์ญ๋์ด ์ฒ๋ฆฌ๋ฉ๋๋ค
|
358 |
+
"""
|
359 |
+
)
|
|
|
360 |
|
361 |
+
return demo
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
362 |
|
363 |
# ์ ํ๋ฆฌ์ผ์ด์
์คํ
|
364 |
if __name__ == "__main__":
|
365 |
+
demo = create_interface()
|
366 |
demo.launch(share=True)
|