Comment by SchemaLoad

13 hours ago

The hosted ones still have the advantage of being able to search the internet for live info rather than being limited to a knowledge cut off date.

I’m not sure why a model needs to be hosted in order to make network calls?

  • Is there a library of good tools for LLMs to call? I have to imagine the bot-detection avoidance mechanisms are a major engineering effort and not likely to work out of the box with a simple harness and random local LLM.

    • If your volume is low enough, it should be pretty fine. It can just piggy back onto your personal browser cookies for Cloudflare.

    • Even the hosted ones are blocked from searching certain sites, for example Claude is banned from searching Reddit:

      `Error: "The following domains are not accessible to our user agent: ['reddit.com']."`

    • Tavily, Exa, Firecrawl, Perplexity, and Linkup are all tools for agents to search the web.

      I’ve been building a harness the past few months and supports them all out of the box with an API key.

      2 replies →

That's not how it works. Whether local or hosted, every modern model has a cutoff date for its training data, and can be leveraged by agents / harnesses / tools to fetch context from the internet or wherever.