Comment by tootyskooty
6 days ago
Yeah it's still hard to deal with LLM gaps (fwiw Study mode would also be prone to this). I do try to catch the super obvious stuff and put up a disclaimer but it's far from perfect.
I had some prototypes basing the generations in websearch but the APIs are still super expensive on that front + the models tend to overindex on the results.
No comments yet
Contribute on Hacker News ↗