Kevin Hu
commited on
Commit
·
7d04573
1
Parent(s):
f586a68
fix minimax bug (#1528)
Browse files### What problem does this PR solve?
#1353
### Type of change
- [x] Bug Fix (non-breaking change which fixes an issue)
- api/db/init_data.py +7 -12
- rag/llm/__init__.py +1 -0
api/db/init_data.py
CHANGED
@@ -573,42 +573,35 @@ def init_llm_factory():
|
|
573 |
# ------------------------ Minimax -----------------------
|
574 |
{
|
575 |
"fid": factory_infos[13]["name"],
|
576 |
-
"llm_name": "abab6.5
|
577 |
"tags": "LLM,CHAT,8k",
|
578 |
"max_tokens": 8192,
|
579 |
"model_type": LLMType.CHAT.value
|
580 |
},
|
581 |
{
|
582 |
"fid": factory_infos[13]["name"],
|
583 |
-
"llm_name": "abab6.5s
|
584 |
"tags": "LLM,CHAT,245k",
|
585 |
"max_tokens": 245760,
|
586 |
"model_type": LLMType.CHAT.value
|
587 |
},
|
588 |
{
|
589 |
"fid": factory_infos[13]["name"],
|
590 |
-
"llm_name": "abab6.5t
|
591 |
"tags": "LLM,CHAT,8k",
|
592 |
"max_tokens": 8192,
|
593 |
"model_type": LLMType.CHAT.value
|
594 |
},
|
595 |
{
|
596 |
"fid": factory_infos[13]["name"],
|
597 |
-
"llm_name": "abab6.5g
|
598 |
"tags": "LLM,CHAT,8k",
|
599 |
"max_tokens": 8192,
|
600 |
"model_type": LLMType.CHAT.value
|
601 |
},
|
602 |
{
|
603 |
"fid": factory_infos[13]["name"],
|
604 |
-
"llm_name": "abab5.
|
605 |
-
"tags": "LLM,CHAT,16k",
|
606 |
-
"max_tokens": 16384,
|
607 |
-
"model_type": LLMType.CHAT.value
|
608 |
-
},
|
609 |
-
{
|
610 |
-
"fid": factory_infos[13]["name"],
|
611 |
-
"llm_name": "abab5.5s-chat",
|
612 |
"tags": "LLM,CHAT,8k",
|
613 |
"max_tokens": 8192,
|
614 |
"model_type": LLMType.CHAT.value
|
@@ -987,6 +980,8 @@ def init_llm_factory():
|
|
987 |
LLMFactoriesService.save(**info)
|
988 |
except Exception as e:
|
989 |
pass
|
|
|
|
|
990 |
for info in llm_infos:
|
991 |
try:
|
992 |
LLMService.save(**info)
|
|
|
573 |
# ------------------------ Minimax -----------------------
|
574 |
{
|
575 |
"fid": factory_infos[13]["name"],
|
576 |
+
"llm_name": "abab6.5",
|
577 |
"tags": "LLM,CHAT,8k",
|
578 |
"max_tokens": 8192,
|
579 |
"model_type": LLMType.CHAT.value
|
580 |
},
|
581 |
{
|
582 |
"fid": factory_infos[13]["name"],
|
583 |
+
"llm_name": "abab6.5s",
|
584 |
"tags": "LLM,CHAT,245k",
|
585 |
"max_tokens": 245760,
|
586 |
"model_type": LLMType.CHAT.value
|
587 |
},
|
588 |
{
|
589 |
"fid": factory_infos[13]["name"],
|
590 |
+
"llm_name": "abab6.5t",
|
591 |
"tags": "LLM,CHAT,8k",
|
592 |
"max_tokens": 8192,
|
593 |
"model_type": LLMType.CHAT.value
|
594 |
},
|
595 |
{
|
596 |
"fid": factory_infos[13]["name"],
|
597 |
+
"llm_name": "abab6.5g",
|
598 |
"tags": "LLM,CHAT,8k",
|
599 |
"max_tokens": 8192,
|
600 |
"model_type": LLMType.CHAT.value
|
601 |
},
|
602 |
{
|
603 |
"fid": factory_infos[13]["name"],
|
604 |
+
"llm_name": "abab5.5s",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
605 |
"tags": "LLM,CHAT,8k",
|
606 |
"max_tokens": 8192,
|
607 |
"model_type": LLMType.CHAT.value
|
|
|
980 |
LLMFactoriesService.save(**info)
|
981 |
except Exception as e:
|
982 |
pass
|
983 |
+
|
984 |
+
LLMService.filter_delete([(LLM.fid == "MiniMax" or LLM.fid == "Minimax")])
|
985 |
for info in llm_infos:
|
986 |
try:
|
987 |
LLMService.save(**info)
|
rag/llm/__init__.py
CHANGED
@@ -61,6 +61,7 @@ ChatModel = {
|
|
61 |
"VolcEngine": VolcEngineChat,
|
62 |
"BaiChuan": BaiChuanChat,
|
63 |
"MiniMax": MiniMaxChat,
|
|
|
64 |
"Mistral": MistralChat,
|
65 |
'Gemini' : GeminiChat,
|
66 |
"Bedrock": BedrockChat,
|
|
|
61 |
"VolcEngine": VolcEngineChat,
|
62 |
"BaiChuan": BaiChuanChat,
|
63 |
"MiniMax": MiniMaxChat,
|
64 |
+
"Minimax": MiniMaxChat,
|
65 |
"Mistral": MistralChat,
|
66 |
'Gemini' : GeminiChat,
|
67 |
"Bedrock": BedrockChat,
|