cecilia-uu
cecilia-uu
commited on
Commit
·
d862dfe
1
Parent(s):
cd3f5c7
Adding the Minimax model (#1009)
Browse files### What problem does this PR solve?
Added support for MiniMax LLM
### Type of change
- [x] New Feature (non-breaking change which adds functionality)
---------
Co-authored-by: cecilia-uu <[email protected]>
- api/db/init_data.py +48 -0
- rag/llm/__init__.py +2 -1
- rag/llm/chat_model.py +8 -0
api/db/init_data.py
CHANGED
@@ -152,6 +152,11 @@ factory_infos = [{
|
|
152 |
"logo": "",
|
153 |
"tags": "TEXT EMBEDDING, TEXT RE-RANK",
|
154 |
"status": "1",
|
|
|
|
|
|
|
|
|
|
|
155 |
}
|
156 |
# {
|
157 |
# "name": "文心一言",
|
@@ -536,6 +541,49 @@ def init_llm_factory():
|
|
536 |
"max_tokens": 2048,
|
537 |
"model_type": LLMType.RERANK.value
|
538 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
539 |
]
|
540 |
for info in factory_infos:
|
541 |
try:
|
|
|
152 |
"logo": "",
|
153 |
"tags": "TEXT EMBEDDING, TEXT RE-RANK",
|
154 |
"status": "1",
|
155 |
+
},{
|
156 |
+
"name": "Minimax",
|
157 |
+
"logo": "",
|
158 |
+
"tags": "LLM,TEXT EMBEDDING",
|
159 |
+
"status": "1",
|
160 |
}
|
161 |
# {
|
162 |
# "name": "文心一言",
|
|
|
541 |
"max_tokens": 2048,
|
542 |
"model_type": LLMType.RERANK.value
|
543 |
},
|
544 |
+
# ------------------------ Minimax -----------------------
|
545 |
+
{
|
546 |
+
"fid": factory_infos[13]["name"],
|
547 |
+
"llm_name": "abab6.5-chat",
|
548 |
+
"tags": "LLM,CHAT,8k",
|
549 |
+
"max_tokens": 8192,
|
550 |
+
"model_type": LLMType.CHAT.value
|
551 |
+
},
|
552 |
+
{
|
553 |
+
"fid": factory_infos[13]["name"],
|
554 |
+
"llm_name": "abab6.5s-chat",
|
555 |
+
"tags": "LLM,CHAT,245k",
|
556 |
+
"max_tokens": 245760,
|
557 |
+
"model_type": LLMType.CHAT.value
|
558 |
+
},
|
559 |
+
{
|
560 |
+
"fid": factory_infos[13]["name"],
|
561 |
+
"llm_name": "abab6.5t-chat",
|
562 |
+
"tags": "LLM,CHAT,8k",
|
563 |
+
"max_tokens": 8192,
|
564 |
+
"model_type": LLMType.CHAT.value
|
565 |
+
},
|
566 |
+
{
|
567 |
+
"fid": factory_infos[13]["name"],
|
568 |
+
"llm_name": "abab6.5g-chat",
|
569 |
+
"tags": "LLM,CHAT,8k",
|
570 |
+
"max_tokens": 8192,
|
571 |
+
"model_type": LLMType.CHAT.value
|
572 |
+
},
|
573 |
+
{
|
574 |
+
"fid": factory_infos[13]["name"],
|
575 |
+
"llm_name": "abab5.5-chat",
|
576 |
+
"tags": "LLM,CHAT,16k",
|
577 |
+
"max_tokens": 16384,
|
578 |
+
"model_type": LLMType.CHAT.value
|
579 |
+
},
|
580 |
+
{
|
581 |
+
"fid": factory_infos[13]["name"],
|
582 |
+
"llm_name": "abab5.5s-chat",
|
583 |
+
"tags": "LLM,CHAT,8k",
|
584 |
+
"max_tokens": 8192,
|
585 |
+
"model_type": LLMType.CHAT.value
|
586 |
+
},
|
587 |
]
|
588 |
for info in factory_infos:
|
589 |
try:
|
rag/llm/__init__.py
CHANGED
@@ -51,7 +51,8 @@ ChatModel = {
|
|
51 |
"Xinference": XinferenceChat,
|
52 |
"Moonshot": MoonshotChat,
|
53 |
"DeepSeek": DeepSeekChat,
|
54 |
-
"BaiChuan": BaiChuanChat
|
|
|
55 |
}
|
56 |
|
57 |
|
|
|
51 |
"Xinference": XinferenceChat,
|
52 |
"Moonshot": MoonshotChat,
|
53 |
"DeepSeek": DeepSeekChat,
|
54 |
+
"BaiChuan": BaiChuanChat,
|
55 |
+
"MiniMax": MiniMaxChat
|
56 |
}
|
57 |
|
58 |
|
rag/llm/chat_model.py
CHANGED
@@ -464,3 +464,11 @@ class VolcEngineChat(Base):
|
|
464 |
except Exception as e:
|
465 |
yield ans + "\n**ERROR**: " + str(e)
|
466 |
yield tk_count
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
464 |
except Exception as e:
|
465 |
yield ans + "\n**ERROR**: " + str(e)
|
466 |
yield tk_count
|
467 |
+
|
468 |
+
|
469 |
+
class MiniMaxChat(Base):
|
470 |
+
def __init__(self, key, model_name="abab6.5s-chat",
|
471 |
+
base_url="https://api.minimax.chat/v1/text/chatcompletion_v2"):
|
472 |
+
if not base_url:
|
473 |
+
base_url="https://api.minimax.chat/v1/text/chatcompletion_v2"
|
474 |
+
super().__init__(key, model_name, base_url)
|