Comment by ah27182
3 days ago
Alright sounds good. Question, what LLM model does this use out of the box? Is it using the models provided by Github (after I give it access)?
3 days ago
Alright sounds good. Question, what LLM model does this use out of the box? Is it using the models provided by Github (after I give it access)?
If you run locally you can mix and match any anthropic / gemini models. As long as it satisfies this protocol https://github.com/appdotbuild/agent/blob/4e0d4b5ac03cee0548... you can plug in anything.
We have a similar wrapper for local LLMs on the roadmap.
If you use CLI only - we run claude 4 + gemini on the backend, gemini serving most of the vision tasks (frontend validation) and claude doing core codegen.
We use both Claude 4 and Gemini by default (for different tasks). But the idea is you can self-host this and use other models (and even BYOM - bring your own models).