Spaces:
Running
Running
Add gemini-2.0-flash-exp LLM
Browse files- app.py +1 -1
- global_config.py +7 -2
app.py
CHANGED
@@ -159,7 +159,7 @@ with st.sidebar:
|
|
159 |
else:
|
160 |
# The LLMs
|
161 |
llm_provider_to_use = st.sidebar.selectbox(
|
162 |
-
label='2: Select
|
163 |
options=[f'{k} ({v["description"]})' for k, v in GlobalConfig.VALID_MODELS.items()],
|
164 |
index=GlobalConfig.DEFAULT_MODEL_INDEX,
|
165 |
help=GlobalConfig.LLM_PROVIDER_HELP,
|
|
|
159 |
else:
|
160 |
# The LLMs
|
161 |
llm_provider_to_use = st.sidebar.selectbox(
|
162 |
+
label='2: Select a suitable LLM to use:\n\n(Gemini and Mistral-Nemo are recommended)',
|
163 |
options=[f'{k} ({v["description"]})' for k, v in GlobalConfig.VALID_MODELS.items()],
|
164 |
index=GlobalConfig.DEFAULT_MODEL_INDEX,
|
165 |
help=GlobalConfig.LLM_PROVIDER_HELP,
|
global_config.py
CHANGED
@@ -34,7 +34,12 @@ class GlobalConfig:
|
|
34 |
'paid': True,
|
35 |
},
|
36 |
'[gg]gemini-1.5-flash-002': {
|
37 |
-
'description': 'faster
|
|
|
|
|
|
|
|
|
|
|
38 |
'max_new_tokens': 8192,
|
39 |
'paid': True,
|
40 |
},
|
@@ -55,7 +60,7 @@ class GlobalConfig:
|
|
55 |
'- **[gg]**: Google Gemini API\n'
|
56 |
'- **[hf]**: Hugging Face Inference API\n'
|
57 |
)
|
58 |
-
DEFAULT_MODEL_INDEX =
|
59 |
LLM_MODEL_TEMPERATURE = 0.2
|
60 |
LLM_MODEL_MIN_OUTPUT_LENGTH = 100
|
61 |
LLM_MODEL_MAX_INPUT_LENGTH = 400 # characters
|
|
|
34 |
'paid': True,
|
35 |
},
|
36 |
'[gg]gemini-1.5-flash-002': {
|
37 |
+
'description': 'faster, detailed',
|
38 |
+
'max_new_tokens': 8192,
|
39 |
+
'paid': True,
|
40 |
+
},
|
41 |
+
'[gg]gemini-2.0-flash-exp': {
|
42 |
+
'description': 'fast, detailed',
|
43 |
'max_new_tokens': 8192,
|
44 |
'paid': True,
|
45 |
},
|
|
|
60 |
'- **[gg]**: Google Gemini API\n'
|
61 |
'- **[hf]**: Hugging Face Inference API\n'
|
62 |
)
|
63 |
+
DEFAULT_MODEL_INDEX = 3
|
64 |
LLM_MODEL_TEMPERATURE = 0.2
|
65 |
LLM_MODEL_MIN_OUTPUT_LENGTH = 100
|
66 |
LLM_MODEL_MAX_INPUT_LENGTH = 400 # characters
|