I have several open source LLM models running on the Hugging Face Inference Endpoint service ( essentially AWS ) …
For all the models I have tested (llama, phi, gemma, etc) … I’m able to connect from Discourse LLM settings page, but inference doesn’t work. here’s the error:
“Trying to contact the model returned this error: Failed to deserialize the JSON body into the target type: missing field inputs at line 1 column 163”
It’s been over a year since I last tried their API. Is it OpenAI compatible nowadays? If so you can try setting Provider to OpenAI and pointing to their endpoint.