Question answering

#39
by codegood - opened

Can this model be used for QA task and if yes, what should the prompt look like?

Is it possible to further train the model (LoRa or Checkpoint) and how? (Sorry if this is a stupid question, I'm still a newbie)

deleted

not an expert but ooba offers in-GUI lora training

Do ooba of them support Mistral?

If I try to run through text generation web, I get this error:

Traceback (most recent call last):
File "E:\ai\ruai\saiga\text-generation-webui\modules\ui_model_menu.py", line 194, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 75, in load_model
output = load_func_maploader
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 136, in huggingface_loader
config = AutoConfig.from_pretrained(path_to_model, trust_remote_code=shared.args.trust_remote_code)
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 1039, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 734, in getitem
raise KeyError(key)
KeyError: 'mistral'

deleted

For what its worth, I can run mistral converted to gguf on ooba just fine on CPU. Cant say about 'regular' model. My GPU is dedicated for other uses.

If I try to run through text generation web, I get this error:

Traceback (most recent call last):
File "E:\ai\ruai\saiga\text-generation-webui\modules\ui_model_menu.py", line 194, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 75, in load_model
output = load_func_maploader
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 136, in huggingface_loader
config = AutoConfig.from_pretrained(path_to_model, trust_remote_code=shared.args.trust_remote_code)
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 1039, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 734, in getitem
raise KeyError(key)
KeyError: 'mistral'

Can you try re-downloading Transformers from Github?

If I try to run through text generation web, I get this error:

Traceback (most recent call last):
File "E:\ai\ruai\saiga\text-generation-webui\modules\ui_model_menu.py", line 194, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 75, in load_model
output = load_func_maploader
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 136, in huggingface_loader
config = AutoConfig.from_pretrained(path_to_model, trust_remote_code=shared.args.trust_remote_code)
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 1039, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 734, in getitem
raise KeyError(key)
KeyError: 'mistral'

Can you try re-downloading Transformers from Github?

I'll try, thanks. (Do you mean that Mistral is using a new version of the transformer? If not, it probably won't work, since Llama starts without problems)

If I try to run through text generation web, I get this error:

Traceback (most recent call last):
File "E:\ai\ruai\saiga\text-generation-webui\modules\ui_model_menu.py", line 194, in load_model_wrapper
shared.model, shared.tokenizer = load_model(shared.model_name, loader)
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 75, in load_model
output = load_func_maploader
File "E:\ai\ruai\saiga\text-generation-webui\modules\models.py", line 136, in huggingface_loader
config = AutoConfig.from_pretrained(path_to_model, trust_remote_code=shared.args.trust_remote_code)
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 1039, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
File "C:\Users\remot.conda\envs\textgen2\lib\site-packages\transformers\models\auto\configuration_auto.py", line 734, in getitem
raise KeyError(key)
KeyError: 'mistral'

Can you try re-downloading Transformers from Github?

I'll try, thanks. (Do you mean that Mistral is using a new version of the transformer? If not, it probably won't work, since Llama starts without problems)
Yeah.

I need to use this model for my question answering project. can I use this directly ? or should I have to train this on a squad dataset for the usage ?

from transformers import AutoTokenizer, AutoModelForQuestionAnswering
model_name = "mistralai/Mixtral-8x7B-Instruct-v0.1"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForQuestionAnswering.from_pretrained(model_name)

@Shashankbr07 You can use it directly

Sign up or log in to comment