Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -211,7 +211,7 @@ async def generate_message_mistral_generate(prompt, max_retries=5):
|
|
211 |
while retries < max_retries:
|
212 |
try:
|
213 |
chat_response = await client_mistral_generate.chat.complete_async(
|
214 |
-
model="mistral-
|
215 |
temperature=1.0,
|
216 |
min_tokens=81,
|
217 |
max_tokens=107,
|
@@ -250,7 +250,7 @@ async def generate_message_mistral_check(prompt, max_retries=5):
|
|
250 |
while retries < max_retries:
|
251 |
try:
|
252 |
chat_response = await client_mistral_check.chat.complete_async(
|
253 |
-
model="mistral-
|
254 |
temperature=0.2,
|
255 |
messages=[
|
256 |
{
|
@@ -287,7 +287,7 @@ async def generate_message_mistral_check_2(prompt, max_retries=5):
|
|
287 |
while retries < max_retries:
|
288 |
try:
|
289 |
chat_response = await client_mistral_check_2.chat.complete_async(
|
290 |
-
model="mistral-
|
291 |
temperature=0.2,
|
292 |
messages=[
|
293 |
{
|
@@ -324,7 +324,7 @@ async def generate_message_mistral_check_3(prompt, max_retries=5):
|
|
324 |
while retries < max_retries:
|
325 |
try:
|
326 |
chat_response = await client_mistral_check_3.chat.complete_async(
|
327 |
-
model="mistral-
|
328 |
temperature=0.2,
|
329 |
messages=[
|
330 |
{
|
@@ -361,7 +361,7 @@ async def generate_message_mistral_check_4(prompt, max_retries=5):
|
|
361 |
while retries < max_retries:
|
362 |
try:
|
363 |
chat_response = await client_mistral_check_4.chat.complete_async(
|
364 |
-
model="mistral-
|
365 |
temperature=0.2,
|
366 |
messages=[
|
367 |
{
|
|
|
211 |
while retries < max_retries:
|
212 |
try:
|
213 |
chat_response = await client_mistral_generate.chat.complete_async(
|
214 |
+
model="open-mistral-nemo",
|
215 |
temperature=1.0,
|
216 |
min_tokens=81,
|
217 |
max_tokens=107,
|
|
|
250 |
while retries < max_retries:
|
251 |
try:
|
252 |
chat_response = await client_mistral_check.chat.complete_async(
|
253 |
+
model="open-mistral-nemo",
|
254 |
temperature=0.2,
|
255 |
messages=[
|
256 |
{
|
|
|
287 |
while retries < max_retries:
|
288 |
try:
|
289 |
chat_response = await client_mistral_check_2.chat.complete_async(
|
290 |
+
model="open-mistral-nemo",
|
291 |
temperature=0.2,
|
292 |
messages=[
|
293 |
{
|
|
|
324 |
while retries < max_retries:
|
325 |
try:
|
326 |
chat_response = await client_mistral_check_3.chat.complete_async(
|
327 |
+
model="open-mistral-nemo",
|
328 |
temperature=0.2,
|
329 |
messages=[
|
330 |
{
|
|
|
361 |
while retries < max_retries:
|
362 |
try:
|
363 |
chat_response = await client_mistral_check_4.chat.complete_async(
|
364 |
+
model="open-mistral-nemo",
|
365 |
temperature=0.2,
|
366 |
messages=[
|
367 |
{
|