Comment by SchemaLoad
17 hours ago
The hosted ones still have the advantage of being able to search the internet for live info rather than being limited to a knowledge cut off date.
17 hours ago
The hosted ones still have the advantage of being able to search the internet for live info rather than being limited to a knowledge cut off date.
I’m not sure why a model needs to be hosted in order to make network calls?
Is there a library of good tools for LLMs to call? I have to imagine the bot-detection avoidance mechanisms are a major engineering effort and not likely to work out of the box with a simple harness and random local LLM.
Even the hosted ones are blocked from searching certain sites, for example Claude is banned from searching Reddit:
`Error: "The following domains are not accessible to our user agent: ['reddit.com']."`
If your volume is low enough, it should be pretty fine. It can just piggy back onto your personal browser cookies for Cloudflare.
Tavily, Exa, Firecrawl, Perplexity, and Linkup are all tools for agents to search the web.
I’ve been building a harness the past few months and supports them all out of the box with an API key.
3 replies →
That's not how it works. Whether local or hosted, every modern model has a cutoff date for its training data, and can be leveraged by agents / harnesses / tools to fetch context from the internet or wherever.
Local ones that support tool use can do the same
You can do that locally too!