Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I guess the parallel is "Ollama serve" which provides you with a direct REST API to interact with a LLM.
 help



llama-cpp provides an API server as well via llama-server (and a competent webgui too).



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: