Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
|
@@ -24,7 +24,7 @@ class MessageRequest(BaseModel):
|
|
| 24 |
message: str
|
| 25 |
language: str
|
| 26 |
|
| 27 |
-
repo_id = "meta-llama/Meta-Llama-3-8B
|
| 28 |
llm_client = InferenceClient(
|
| 29 |
model=repo_id,
|
| 30 |
token=os.getenv("HF_TOKEN"),
|
|
@@ -60,8 +60,8 @@ templates = Jinja2Templates(directory="static")
|
|
| 60 |
|
| 61 |
# Configure Llama index settings
|
| 62 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 63 |
-
model_name="meta-llama/Meta-Llama-3-8B
|
| 64 |
-
tokenizer_name="meta-llama/Meta-Llama-3-8B
|
| 65 |
context_window=3000,
|
| 66 |
token=os.getenv("HF_TOKEN"),
|
| 67 |
max_new_tokens=512,
|
|
|
|
| 24 |
message: str
|
| 25 |
language: str
|
| 26 |
|
| 27 |
+
repo_id = "meta-llama/Meta-Llama-3-8B"
|
| 28 |
llm_client = InferenceClient(
|
| 29 |
model=repo_id,
|
| 30 |
token=os.getenv("HF_TOKEN"),
|
|
|
|
| 60 |
|
| 61 |
# Configure Llama index settings
|
| 62 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 63 |
+
model_name="meta-llama/Meta-Llama-3-8B",
|
| 64 |
+
tokenizer_name="meta-llama/Meta-Llama-3-8B",
|
| 65 |
context_window=3000,
|
| 66 |
token=os.getenv("HF_TOKEN"),
|
| 67 |
max_new_tokens=512,
|