← Back to context

Comment by lxgr

7 days ago

Can it easily run as a server process in the background? To me, not having to load the LLM into memory for every single interaction is a big win of Ollama.

Yes, of course it can.

  • I wouldn't consider that a given at all, but apparently there's indeed `llama-server` which looks promising!

    Then the only thing that's missing seems to be a canonical way for clients to instantiate that, ideally in some OS-native way (systemd, launchcd etc.), and a canonical port that they can connect to.