Spaces:
Running
Suggestion - provide API for current models with endpoint for chat (as the huggingchat llama3-70b version is superior to others)
Hello,
I've been comparing llama3-70b from huggingchat, groq, deepinfra (via API) and meta.ai and the huggingchat instance is superior to the others.
Is it possible for huggingface to provide a chat API that we can use (with a completion etc) instead of having to run the models ourselves (which appears to be the case at the moment? I have a token key but seems one needs to run the model to use it?).
Alternatively, does anyone know the system instructions that causes the huggingchat instances to give such good responses?
Thanks
It might be that HuggingChat has no system instructions by default, and that makes it better. All I know is that meta.ai has a system prompt causing it to act weird, and is much worse than HuggingChat
Sure, maybe.
If we could get access to it via API, paid or not that would be great!
Ah, really? So you don't have to run anything?
That's good, though I was looking for typical API pricing (per token rather than monthly).