Spaces:
Running
Running
Update app.py
Browse filesFixed chat changed state var model after image analysis chat
app.py
CHANGED
@@ -176,6 +176,7 @@ def updatePassword(txt):
|
|
176 |
# return val
|
177 |
|
178 |
def chat(prompt, user_window, pwd_window, past, response, gptModel, uploaded_image_file=''):
|
|
|
179 |
user_window = user_window.lower().strip()
|
180 |
isBoss = False
|
181 |
if user_window == unames[0] and pwd_window == pwdList[0]:
|
@@ -200,13 +201,15 @@ def chat(prompt, user_window, pwd_window, past, response, gptModel, uploaded_ima
|
|
200 |
return [past, response, None, gptModel, uploaded_image_file]
|
201 |
if user_window in unames and pwd_window == pwdList[unames.index(user_window)]:
|
202 |
past.append({"role":"user", "content":prompt})
|
203 |
-
|
|
|
204 |
completion = client.chat.completions.create(model=gptModel,
|
205 |
messages=past)
|
|
|
206 |
else:
|
207 |
-
|
208 |
-
(completion, msg) = analyze_image(user_window, gptModel)
|
209 |
uploaded_image_file= ''
|
|
|
210 |
if not msg == 'ok':
|
211 |
return [past, msg, None, gptModel, uploaded_image_file]
|
212 |
reply = completion.choices[0].message.content
|
@@ -215,7 +218,7 @@ def chat(prompt, user_window, pwd_window, past, response, gptModel, uploaded_ima
|
|
215 |
tokens = completion.usage.total_tokens
|
216 |
response += "\n\nYOU: " + prompt + "\nGPT: " + reply
|
217 |
if isBoss:
|
218 |
-
response += f"\n{
|
219 |
if tokens > 40000:
|
220 |
response += "\n\nTHIS DIALOG IS GETTING TOO LONG. PLEASE RESTART CONVERSATION SOON."
|
221 |
past.append({"role":"assistant", "content": reply})
|
@@ -225,7 +228,7 @@ def chat(prompt, user_window, pwd_window, past, response, gptModel, uploaded_ima
|
|
225 |
dataFile = new_func(user_window)
|
226 |
with open(dataFile, 'a') as f:
|
227 |
m = '4o'
|
228 |
-
if 'mini' in
|
229 |
m = '4omini'
|
230 |
f.write(f'{user_window}:{tokens_in}/{tokens_out}-{m}\n')
|
231 |
accessOk = True
|
|
|
176 |
# return val
|
177 |
|
178 |
def chat(prompt, user_window, pwd_window, past, response, gptModel, uploaded_image_file=''):
|
179 |
+
image_gen_model = 'gpt-4o-2024-08-06'
|
180 |
user_window = user_window.lower().strip()
|
181 |
isBoss = False
|
182 |
if user_window == unames[0] and pwd_window == pwdList[0]:
|
|
|
201 |
return [past, response, None, gptModel, uploaded_image_file]
|
202 |
if user_window in unames and pwd_window == pwdList[unames.index(user_window)]:
|
203 |
past.append({"role":"user", "content":prompt})
|
204 |
+
gen_image = (uploaded_image_file != '')
|
205 |
+
if not gen_image:
|
206 |
completion = client.chat.completions.create(model=gptModel,
|
207 |
messages=past)
|
208 |
+
reporting_model = gptModel
|
209 |
else:
|
210 |
+
(completion, msg) = analyze_image(user_window, image_gen_model)
|
|
|
211 |
uploaded_image_file= ''
|
212 |
+
reporting_model = image_gen_model
|
213 |
if not msg == 'ok':
|
214 |
return [past, msg, None, gptModel, uploaded_image_file]
|
215 |
reply = completion.choices[0].message.content
|
|
|
218 |
tokens = completion.usage.total_tokens
|
219 |
response += "\n\nYOU: " + prompt + "\nGPT: " + reply
|
220 |
if isBoss:
|
221 |
+
response += f"\n{reporting_model}: tokens in/out = {tokens_in}/{tokens_out}"
|
222 |
if tokens > 40000:
|
223 |
response += "\n\nTHIS DIALOG IS GETTING TOO LONG. PLEASE RESTART CONVERSATION SOON."
|
224 |
past.append({"role":"assistant", "content": reply})
|
|
|
228 |
dataFile = new_func(user_window)
|
229 |
with open(dataFile, 'a') as f:
|
230 |
m = '4o'
|
231 |
+
if 'mini' in reporting_model:
|
232 |
m = '4omini'
|
233 |
f.write(f'{user_window}:{tokens_in}/{tokens_out}-{m}\n')
|
234 |
accessOk = True
|