Spaces:
Sleeping
Sleeping
Update yourbench_space/leaderboard_space/env.py
Browse files
yourbench_space/leaderboard_space/env.py
CHANGED
@@ -3,19 +3,16 @@ import os
|
|
3 |
|
4 |
INIT_MODELS = [
|
5 |
# ("meta-llama/Llama-4-Scout-17B-16E-Instruct", "together"),
|
6 |
-
("Qwen/Qwen3-32B", "
|
7 |
-
("Qwen/Qwen2.5-72B-Instruct", "
|
8 |
-
("
|
9 |
-
("deepseek-ai/DeepSeek-
|
10 |
-
(
|
11 |
-
(
|
12 |
-
(
|
13 |
-
('
|
14 |
-
('
|
15 |
-
|
16 |
-
('google/gemma-2-27b-it','nebius'),
|
17 |
-
('google/gemma-2-9b-it','nebius'),
|
18 |
-
('google/gemma-2-2b-it','nebius'),
|
19 |
]
|
20 |
MODELS = [m[0] for m in INIT_MODELS]
|
21 |
TASK = os.getenv("TASK")
|
|
|
3 |
|
4 |
INIT_MODELS = [
|
5 |
# ("meta-llama/Llama-4-Scout-17B-16E-Instruct", "together"),
|
6 |
+
("Qwen/Qwen3-32B", "hf-inference"),
|
7 |
+
("Qwen/Qwen2.5-72B-Instruct", "hf-inference"),
|
8 |
+
("Qwen/Qwen3-235B-A22B", "hf-inference"),
|
9 |
+
("deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "hf-inference"),
|
10 |
+
('meta-llama/Llama-3.1-8B-Instruct','hf-inference'),
|
11 |
+
('meta-llama/Llama-3.3-70B-Instruct','hf-inference'),
|
12 |
+
('microsoft/phi-4','hf-inference'),
|
13 |
+
('microsoft/Phi-3-mini-4k-instruct','hf-inference'),
|
14 |
+
('CohereLabs/c4ai-command-r-plus-08-2024','hf-inference'),
|
15 |
+
|
|
|
|
|
|
|
16 |
]
|
17 |
MODELS = [m[0] for m in INIT_MODELS]
|
18 |
TASK = os.getenv("TASK")
|