Spaces:
Running
Running
future-xy
commited on
Commit
·
3020792
1
Parent(s):
5fd4d0a
add base class
Browse files
backend-cli.py
CHANGED
|
@@ -406,8 +406,8 @@ if __name__ == "__main__":
|
|
| 406 |
if local_debug:
|
| 407 |
debug_model_names = ["mistralai/Mixtral-8x7B-Instruct-v0.1"]
|
| 408 |
# debug_model_names = ["TheBloke/Mixtral-8x7B-v0.1-GPTQ"]
|
| 409 |
-
|
| 410 |
-
debug_task_name = "mmlu"
|
| 411 |
task_lst = TASKS_HARNESS.copy()
|
| 412 |
for task in task_lst:
|
| 413 |
for debug_model_name in debug_model_names:
|
|
|
|
| 406 |
if local_debug:
|
| 407 |
debug_model_names = ["mistralai/Mixtral-8x7B-Instruct-v0.1"]
|
| 408 |
# debug_model_names = ["TheBloke/Mixtral-8x7B-v0.1-GPTQ"]
|
| 409 |
+
debug_task_name = 'selfcheckgpt'
|
| 410 |
+
# debug_task_name = "mmlu"
|
| 411 |
task_lst = TASKS_HARNESS.copy()
|
| 412 |
for task in task_lst:
|
| 413 |
for debug_model_name in debug_model_names:
|
src/backend/huggingface_generate_until.py
CHANGED
|
@@ -2,12 +2,13 @@ from typing import List, Literal, Optional, Tuple, Union
|
|
| 2 |
import torch
|
| 3 |
import transformers
|
| 4 |
|
| 5 |
-
from lm_eval.models.huggingface import HFLM
|
| 6 |
from lm_eval.api.registry import register_model
|
| 7 |
|
|
|
|
|
|
|
| 8 |
|
| 9 |
@register_model("hf-chat")
|
| 10 |
-
class HFLMwithChatTemplate(
|
| 11 |
def __init__(self, use_chat_template=True, **kwargs):
|
| 12 |
super().__init__(**kwargs)
|
| 13 |
self.use_chat_template = use_chat_template
|
|
|
|
| 2 |
import torch
|
| 3 |
import transformers
|
| 4 |
|
|
|
|
| 5 |
from lm_eval.api.registry import register_model
|
| 6 |
|
| 7 |
+
from src.backend.hflm_with_measurement import HFLMWithMeasurement
|
| 8 |
+
|
| 9 |
|
| 10 |
@register_model("hf-chat")
|
| 11 |
+
class HFLMwithChatTemplate(HFLMWithMeasurement):
|
| 12 |
def __init__(self, use_chat_template=True, **kwargs):
|
| 13 |
super().__init__(**kwargs)
|
| 14 |
self.use_chat_template = use_chat_template
|
src/backend/moe_infinity.py
CHANGED
|
@@ -5,12 +5,13 @@ from transformers import AutoModelForCausalLM
|
|
| 5 |
from moe_infinity import MoE
|
| 6 |
from typing import List, Tuple, Optional, Union
|
| 7 |
|
| 8 |
-
from lm_eval.models.huggingface import HFLM
|
| 9 |
from lm_eval.api.registry import register_model
|
| 10 |
|
|
|
|
|
|
|
| 11 |
|
| 12 |
@register_model("moe-infinity")
|
| 13 |
-
class MoEHFLM(
|
| 14 |
def __init__(
|
| 15 |
self,
|
| 16 |
pretrained: str = "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
|
|
|
| 5 |
from moe_infinity import MoE
|
| 6 |
from typing import List, Tuple, Optional, Union
|
| 7 |
|
|
|
|
| 8 |
from lm_eval.api.registry import register_model
|
| 9 |
|
| 10 |
+
from src.backend.hflm_with_measurement import HFLMWithMeasurement
|
| 11 |
+
|
| 12 |
|
| 13 |
@register_model("moe-infinity")
|
| 14 |
+
class MoEHFLM(HFLMWithMeasurement):
|
| 15 |
def __init__(
|
| 16 |
self,
|
| 17 |
pretrained: str = "mistralai/Mixtral-8x7B-Instruct-v0.1",
|