Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -99,7 +99,8 @@ def respond_deepseek_qna(
|
|
99 |
system_message: str,
|
100 |
max_tokens: int,
|
101 |
temperature: float,
|
102 |
-
top_p: float
|
|
|
103 |
):
|
104 |
"""
|
105 |
DeepSeek 모델을 이용해 한 번의 질문(question)에 대한 답변을 반환하는 함수.
|
@@ -118,7 +119,7 @@ def respond_deepseek_qna(
|
|
118 |
|
119 |
try:
|
120 |
response = openai.ChatCompletion.create(
|
121 |
-
model=
|
122 |
messages=messages,
|
123 |
max_tokens=max_tokens,
|
124 |
temperature=temperature,
|
@@ -344,6 +345,13 @@ with gr.Blocks() as demo:
|
|
344 |
# DeepSeek 탭
|
345 |
#################
|
346 |
with gr.Tab("DeepSeek-V3"):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
347 |
deepseek_input1 = gr.Textbox(label="입력1", lines=1)
|
348 |
deepseek_input2 = gr.Textbox(label="입력2", lines=1)
|
349 |
deepseek_input3 = gr.Textbox(label="입력3", lines=1)
|
@@ -367,14 +375,21 @@ with gr.Blocks() as demo:
|
|
367 |
|
368 |
deepseek_submit_button = gr.Button("전송")
|
369 |
|
370 |
-
def merge_and_call_deepseek(i1, i2, i3, i4, i5, sys_msg, mt, temp, top_p_):
|
|
|
|
|
|
|
|
|
|
|
|
|
371 |
question = " ".join([i1, i2, i3, i4, i5])
|
372 |
return respond_deepseek_qna(
|
373 |
question=question,
|
374 |
system_message=sys_msg,
|
375 |
max_tokens=mt,
|
376 |
temperature=temp,
|
377 |
-
top_p=top_p_
|
|
|
378 |
)
|
379 |
|
380 |
deepseek_submit_button.click(
|
@@ -384,7 +399,8 @@ with gr.Blocks() as demo:
|
|
384 |
deepseek_system_message,
|
385 |
deepseek_max_tokens,
|
386 |
deepseek_temperature,
|
387 |
-
deepseek_top_p
|
|
|
388 |
],
|
389 |
outputs=deepseek_answer_output
|
390 |
)
|
@@ -393,4 +409,4 @@ with gr.Blocks() as demo:
|
|
393 |
# 메인 실행부
|
394 |
#############################
|
395 |
if __name__ == "__main__":
|
396 |
-
demo.launch()
|
|
|
99 |
system_message: str,
|
100 |
max_tokens: int,
|
101 |
temperature: float,
|
102 |
+
top_p: float,
|
103 |
+
model_name: str # 모델 이름 추가
|
104 |
):
|
105 |
"""
|
106 |
DeepSeek 모델을 이용해 한 번의 질문(question)에 대한 답변을 반환하는 함수.
|
|
|
119 |
|
120 |
try:
|
121 |
response = openai.ChatCompletion.create(
|
122 |
+
model=model_name, # 선택된 모델 사용
|
123 |
messages=messages,
|
124 |
max_tokens=max_tokens,
|
125 |
temperature=temperature,
|
|
|
345 |
# DeepSeek 탭
|
346 |
#################
|
347 |
with gr.Tab("DeepSeek-V3"):
|
348 |
+
# 라디오 버튼 추가
|
349 |
+
deepseek_model_radio = gr.Radio(
|
350 |
+
choices=["V3 (deepseek-chat)", "R1 (deepseek-reasoner)"], # 선택지
|
351 |
+
label="모델 선택", # 라벨
|
352 |
+
value="V3 (deepseek-chat)" # 기본값
|
353 |
+
)
|
354 |
+
|
355 |
deepseek_input1 = gr.Textbox(label="입력1", lines=1)
|
356 |
deepseek_input2 = gr.Textbox(label="입력2", lines=1)
|
357 |
deepseek_input3 = gr.Textbox(label="입력3", lines=1)
|
|
|
375 |
|
376 |
deepseek_submit_button = gr.Button("전송")
|
377 |
|
378 |
+
def merge_and_call_deepseek(i1, i2, i3, i4, i5, sys_msg, mt, temp, top_p_, model_radio):
|
379 |
+
# 라디오 버튼에서 선택된 모델 이름 추출
|
380 |
+
if model_radio == "V3 (deepseek-chat)":
|
381 |
+
model_name = "deepseek-chat"
|
382 |
+
else:
|
383 |
+
model_name = "deepseek-reasoner"
|
384 |
+
|
385 |
question = " ".join([i1, i2, i3, i4, i5])
|
386 |
return respond_deepseek_qna(
|
387 |
question=question,
|
388 |
system_message=sys_msg,
|
389 |
max_tokens=mt,
|
390 |
temperature=temp,
|
391 |
+
top_p=top_p_,
|
392 |
+
model_name=model_name # 선택된 모델 이름 전달
|
393 |
)
|
394 |
|
395 |
deepseek_submit_button.click(
|
|
|
399 |
deepseek_system_message,
|
400 |
deepseek_max_tokens,
|
401 |
deepseek_temperature,
|
402 |
+
deepseek_top_p,
|
403 |
+
deepseek_model_radio # 라디오 버튼 입력 추가
|
404 |
],
|
405 |
outputs=deepseek_answer_output
|
406 |
)
|
|
|
409 |
# 메인 실행부
|
410 |
#############################
|
411 |
if __name__ == "__main__":
|
412 |
+
demo.launch()
|