inference-endpoints
this will allow deployment on huggingface inference endpoint infrastructure
Thanks for making the model available for deployment!
I have added my own inference endpoint using your forked repo, but when I type a prompt in the "Test your endpoint!" text box, I get an error:
API Implementation Error: Invalid output: output must be of type <conversation: <generated_responses:Array; past_user_inputs:Array>>
Any idea what I'm doing wrong?
Thanks!
Much appreciated if you could provide inference. We would like to evaluate the model for our use cases, would be happy to share our insights too!
Thanks for making the model available for deployment!
I have added my own inference endpoint using your forked repo, but when I type a prompt in the "Test your endpoint!" text box, I get an error:
API Implementation Error: Invalid output: output must be of type <conversation: <generated_responses:Array; past_user_inputs:Array>>Any idea what I'm doing wrong?
Thanks!
Hi @infojunkie
once you have it deployed - use curl or python to call the api
i have added some desc here
https://huggingface.co/poiccard/jais-13b-chat-adn
Thank you @poiccard for your contribution.