Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -209,12 +209,12 @@ async def generate_message_mistral_generate(prompt, max_retries=5):
|
|
209 |
chat_response = await client.chat.completions.create(
|
210 |
model="gpt-4o-mini",
|
211 |
messages=[{"role": "user", "content": prompt}],
|
212 |
-
temperature=1.
|
213 |
max_tokens=101
|
214 |
)
|
215 |
|
216 |
-
#
|
217 |
-
cleaned_message = clean_message(chat_response.choices[0].message
|
218 |
return cleaned_message
|
219 |
except Exception as e:
|
220 |
error_message = str(e)
|
@@ -228,6 +228,7 @@ async def generate_message_mistral_generate(prompt, max_retries=5):
|
|
228 |
return None
|
229 |
print("Failed to get response from GPT-4o-mini after maximum retries.")
|
230 |
return None
|
|
|
231 |
#def generate_message_mistral_generate(prompt):
|
232 |
# try:
|
233 |
# messages = [SystemMessage(content=prompt)]
|
|
|
209 |
chat_response = await client.chat.completions.create(
|
210 |
model="gpt-4o-mini",
|
211 |
messages=[{"role": "user", "content": prompt}],
|
212 |
+
temperature=1.0,
|
213 |
max_tokens=101
|
214 |
)
|
215 |
|
216 |
+
# Properly access the message content
|
217 |
+
cleaned_message = clean_message(chat_response.choices[0].message.content.strip())
|
218 |
return cleaned_message
|
219 |
except Exception as e:
|
220 |
error_message = str(e)
|
|
|
228 |
return None
|
229 |
print("Failed to get response from GPT-4o-mini after maximum retries.")
|
230 |
return None
|
231 |
+
|
232 |
#def generate_message_mistral_generate(prompt):
|
233 |
# try:
|
234 |
# messages = [SystemMessage(content=prompt)]
|