I'm using this llama.cpp docker image "ghcr.io/ggerganov/llama.cpp:server", but i can't infer model. Do you have format prompt?
· Sign up or log in to comment