Spaces:
Building
Building
Update utils/llms.py
Browse files- utils/llms.py +7 -7
utils/llms.py
CHANGED
@@ -6,7 +6,7 @@ from g4f.client import Client
|
|
6 |
from litellm import completion
|
7 |
import random
|
8 |
|
9 |
-
from g4f.Provider import DeepInfraChat,Glider,Liaobots,Blackbox,ChatGptEs
|
10 |
os.environ["OPENROUTER_API_KEY"] = "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"
|
11 |
os.environ["GROQ_API_KEY"] ="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"
|
12 |
gemini_api_keys=["AIzaSyB7yKIdfW7Umv62G47BCdJjoHTJ9TeiAko","AIzaSyDtP05TyoIy9j0uPL7_wLEhgQEE75AZQSc","AIzaSyDOyjfqFhHmGlGJ2raX82XWTtmMcZxRshs"]
|
@@ -14,12 +14,12 @@ groq_api_keys=["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","gsk_b
|
|
14 |
#[,"AIzaSyBPfR-HG_HeUgLF0LYW1XQgQUxFF6jF_0U","AIzaSyBz01gZCb9kzZF3lNHuwy_iajWhi9ivyDk"]]
|
15 |
os.environ["GEMINI_API_KEY"] =random.choice(gemini_api_keys)
|
16 |
os.environ["TOGETHERAI_API_KEY"] ="30bed0b842ed3268372d57f588c22452798e9af96aa8d3129ba745ef226282a8"
|
17 |
-
DeepInfraChat.models = ["deepseek-ai/DeepSeek-R1-Turbo","Qwen/QwQ-32B","deepseek-ai/DeepSeek-R1"]
|
18 |
|
19 |
-
REASONING_CORRESPONDANCE = {"DeepSeek-R1-Glider":Glider,"DeepSeekR1":DeepInfraChat}
|
20 |
REASONING_QWQ = {"qwq-32b":DeepInfraChat}
|
21 |
|
22 |
-
CHAT_CORRESPONDANCE = {"DeepSeek-V3":DeepInfraChat
|
23 |
|
24 |
client = Client()
|
25 |
genai.configure(api_key="AIzaSyAQgAtQPpY0bQaCqCISGxeyF6tpDePx-Jg")
|
@@ -137,12 +137,12 @@ def gpt4stream(messages,model,api_keys):
|
|
137 |
helper.q.put_nowait("RESULT: "+cunk)
|
138 |
|
139 |
|
140 |
-
elif model=="deepseek.r1":
|
141 |
cunk=""
|
142 |
|
143 |
-
if "
|
144 |
providers = CHAT_CORRESPONDANCE
|
145 |
-
model_name="deepseek-
|
146 |
else:
|
147 |
providers = REASONING_CORRESPONDANCE
|
148 |
model_name="deepseek-r1"
|
|
|
6 |
from litellm import completion
|
7 |
import random
|
8 |
|
9 |
+
from g4f.Provider import DeepInfraChat,Glider,Liaobots,Blackbox,ChatGptEs,LambdaChat,TypeGPT
|
10 |
os.environ["OPENROUTER_API_KEY"] = "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"
|
11 |
os.environ["GROQ_API_KEY"] ="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"
|
12 |
gemini_api_keys=["AIzaSyB7yKIdfW7Umv62G47BCdJjoHTJ9TeiAko","AIzaSyDtP05TyoIy9j0uPL7_wLEhgQEE75AZQSc","AIzaSyDOyjfqFhHmGlGJ2raX82XWTtmMcZxRshs"]
|
|
|
14 |
#[,"AIzaSyBPfR-HG_HeUgLF0LYW1XQgQUxFF6jF_0U","AIzaSyBz01gZCb9kzZF3lNHuwy_iajWhi9ivyDk"]]
|
15 |
os.environ["GEMINI_API_KEY"] =random.choice(gemini_api_keys)
|
16 |
os.environ["TOGETHERAI_API_KEY"] ="30bed0b842ed3268372d57f588c22452798e9af96aa8d3129ba745ef226282a8"
|
17 |
+
DeepInfraChat.models = ["deepseek-ai/DeepSeek-R1-Turbo","Qwen/QwQ-32B","deepseek-ai/DeepSeek-R1","deepseek-ai/DeepSeek-V3-0324"]
|
18 |
|
19 |
+
REASONING_CORRESPONDANCE = {"DeepSeek-R1-Glider":Glider, "DeepSeekR1-LAMBDA":LambdaChat,"DeepSeekR1":DeepInfraChat,"deepseek-slow":TypeGPT}
|
20 |
REASONING_QWQ = {"qwq-32b":DeepInfraChat}
|
21 |
|
22 |
+
CHAT_CORRESPONDANCE = {"DeepSeek-V3":DeepInfraChat}
|
23 |
|
24 |
client = Client()
|
25 |
genai.configure(api_key="AIzaSyAQgAtQPpY0bQaCqCISGxeyF6tpDePx-Jg")
|
|
|
137 |
helper.q.put_nowait("RESULT: "+cunk)
|
138 |
|
139 |
|
140 |
+
elif model=="deepseek.r1" or model=="deepseek-chat":
|
141 |
cunk=""
|
142 |
|
143 |
+
if "chat" in model:
|
144 |
providers = CHAT_CORRESPONDANCE
|
145 |
+
model_name="deepseek-ai/DeepSeek-V3-0324"
|
146 |
else:
|
147 |
providers = REASONING_CORRESPONDANCE
|
148 |
model_name="deepseek-r1"
|