fix LocalAI add bug (#1851)
Browse files### What problem does this PR solve?
#1848
### Type of change
- [x] Bug Fix (non-breaking change which fixes an issue)
---------
Co-authored-by: Zhedong Cen <[email protected]>
- rag/llm/chat_model.py +3 -3
- rag/llm/embedding_model.py +3 -3
rag/llm/chat_model.py
CHANGED
|
@@ -102,7 +102,7 @@ class XinferenceChat(Base):
|
|
| 102 |
if not base_url:
|
| 103 |
raise ValueError("Local llm url cannot be None")
|
| 104 |
if base_url.split("/")[-1] != "v1":
|
| 105 |
-
|
| 106 |
key = "xxx"
|
| 107 |
super().__init__(key, model_name, base_url)
|
| 108 |
|
|
@@ -373,8 +373,8 @@ class LocalAIChat(Base):
|
|
| 373 |
if not base_url:
|
| 374 |
raise ValueError("Local llm url cannot be None")
|
| 375 |
if base_url.split("/")[-1] != "v1":
|
| 376 |
-
|
| 377 |
-
self.client = OpenAI(api_key="empty", base_url=
|
| 378 |
self.model_name = model_name.split("___")[0]
|
| 379 |
|
| 380 |
|
|
|
|
| 102 |
if not base_url:
|
| 103 |
raise ValueError("Local llm url cannot be None")
|
| 104 |
if base_url.split("/")[-1] != "v1":
|
| 105 |
+
base_url = os.path.join(base_url, "v1")
|
| 106 |
key = "xxx"
|
| 107 |
super().__init__(key, model_name, base_url)
|
| 108 |
|
|
|
|
| 373 |
if not base_url:
|
| 374 |
raise ValueError("Local llm url cannot be None")
|
| 375 |
if base_url.split("/")[-1] != "v1":
|
| 376 |
+
base_url = os.path.join(base_url, "v1")
|
| 377 |
+
self.client = OpenAI(api_key="empty", base_url=base_url)
|
| 378 |
self.model_name = model_name.split("___")[0]
|
| 379 |
|
| 380 |
|
rag/llm/embedding_model.py
CHANGED
|
@@ -510,8 +510,8 @@ class LmStudioEmbed(LocalAIEmbed):
|
|
| 510 |
if not base_url:
|
| 511 |
raise ValueError("Local llm url cannot be None")
|
| 512 |
if base_url.split("/")[-1] != "v1":
|
| 513 |
-
|
| 514 |
-
self.client = OpenAI(api_key="lm-studio", base_url=
|
| 515 |
self.model_name = model_name
|
| 516 |
|
| 517 |
|
|
@@ -520,6 +520,6 @@ class OpenAI_APIEmbed(OpenAIEmbed):
|
|
| 520 |
if not base_url:
|
| 521 |
raise ValueError("url cannot be None")
|
| 522 |
if base_url.split("/")[-1] != "v1":
|
| 523 |
-
|
| 524 |
self.client = OpenAI(api_key=key, base_url=base_url)
|
| 525 |
self.model_name = model_name.split("___")[0]
|
|
|
|
| 510 |
if not base_url:
|
| 511 |
raise ValueError("Local llm url cannot be None")
|
| 512 |
if base_url.split("/")[-1] != "v1":
|
| 513 |
+
base_url = os.path.join(base_url, "v1")
|
| 514 |
+
self.client = OpenAI(api_key="lm-studio", base_url=base_url)
|
| 515 |
self.model_name = model_name
|
| 516 |
|
| 517 |
|
|
|
|
| 520 |
if not base_url:
|
| 521 |
raise ValueError("url cannot be None")
|
| 522 |
if base_url.split("/")[-1] != "v1":
|
| 523 |
+
base_url = os.path.join(base_url, "v1")
|
| 524 |
self.client = OpenAI(api_key=key, base_url=base_url)
|
| 525 |
self.model_name = model_name.split("___")[0]
|