cecilia-uu cecilia-uu commited on
Commit
d862dfe
·
1 Parent(s): cd3f5c7

Adding the Minimax model (#1009)

Browse files

### What problem does this PR solve?

Added support for MiniMax LLM

### Type of change

- [x] New Feature (non-breaking change which adds functionality)

---------

Co-authored-by: cecilia-uu <[email protected]>

api/db/init_data.py CHANGED
@@ -152,6 +152,11 @@ factory_infos = [{
152
  "logo": "",
153
  "tags": "TEXT EMBEDDING, TEXT RE-RANK",
154
  "status": "1",
 
 
 
 
 
155
  }
156
  # {
157
  # "name": "文心一言",
@@ -536,6 +541,49 @@ def init_llm_factory():
536
  "max_tokens": 2048,
537
  "model_type": LLMType.RERANK.value
538
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
539
  ]
540
  for info in factory_infos:
541
  try:
 
152
  "logo": "",
153
  "tags": "TEXT EMBEDDING, TEXT RE-RANK",
154
  "status": "1",
155
+ },{
156
+ "name": "Minimax",
157
+ "logo": "",
158
+ "tags": "LLM,TEXT EMBEDDING",
159
+ "status": "1",
160
  }
161
  # {
162
  # "name": "文心一言",
 
541
  "max_tokens": 2048,
542
  "model_type": LLMType.RERANK.value
543
  },
544
+ # ------------------------ Minimax -----------------------
545
+ {
546
+ "fid": factory_infos[13]["name"],
547
+ "llm_name": "abab6.5-chat",
548
+ "tags": "LLM,CHAT,8k",
549
+ "max_tokens": 8192,
550
+ "model_type": LLMType.CHAT.value
551
+ },
552
+ {
553
+ "fid": factory_infos[13]["name"],
554
+ "llm_name": "abab6.5s-chat",
555
+ "tags": "LLM,CHAT,245k",
556
+ "max_tokens": 245760,
557
+ "model_type": LLMType.CHAT.value
558
+ },
559
+ {
560
+ "fid": factory_infos[13]["name"],
561
+ "llm_name": "abab6.5t-chat",
562
+ "tags": "LLM,CHAT,8k",
563
+ "max_tokens": 8192,
564
+ "model_type": LLMType.CHAT.value
565
+ },
566
+ {
567
+ "fid": factory_infos[13]["name"],
568
+ "llm_name": "abab6.5g-chat",
569
+ "tags": "LLM,CHAT,8k",
570
+ "max_tokens": 8192,
571
+ "model_type": LLMType.CHAT.value
572
+ },
573
+ {
574
+ "fid": factory_infos[13]["name"],
575
+ "llm_name": "abab5.5-chat",
576
+ "tags": "LLM,CHAT,16k",
577
+ "max_tokens": 16384,
578
+ "model_type": LLMType.CHAT.value
579
+ },
580
+ {
581
+ "fid": factory_infos[13]["name"],
582
+ "llm_name": "abab5.5s-chat",
583
+ "tags": "LLM,CHAT,8k",
584
+ "max_tokens": 8192,
585
+ "model_type": LLMType.CHAT.value
586
+ },
587
  ]
588
  for info in factory_infos:
589
  try:
rag/llm/__init__.py CHANGED
@@ -51,7 +51,8 @@ ChatModel = {
51
  "Xinference": XinferenceChat,
52
  "Moonshot": MoonshotChat,
53
  "DeepSeek": DeepSeekChat,
54
- "BaiChuan": BaiChuanChat
 
55
  }
56
 
57
 
 
51
  "Xinference": XinferenceChat,
52
  "Moonshot": MoonshotChat,
53
  "DeepSeek": DeepSeekChat,
54
+ "BaiChuan": BaiChuanChat,
55
+ "MiniMax": MiniMaxChat
56
  }
57
 
58
 
rag/llm/chat_model.py CHANGED
@@ -464,3 +464,11 @@ class VolcEngineChat(Base):
464
  except Exception as e:
465
  yield ans + "\n**ERROR**: " + str(e)
466
  yield tk_count
 
 
 
 
 
 
 
 
 
464
  except Exception as e:
465
  yield ans + "\n**ERROR**: " + str(e)
466
  yield tk_count
467
+
468
+
469
+ class MiniMaxChat(Base):
470
+ def __init__(self, key, model_name="abab6.5s-chat",
471
+ base_url="https://api.minimax.chat/v1/text/chatcompletion_v2"):
472
+ if not base_url:
473
+ base_url="https://api.minimax.chat/v1/text/chatcompletion_v2"
474
+ super().__init__(key, model_name, base_url)