← Back to context Comment by __loam 7 months ago This is why hallucinations will never be fixed in language models. That's just how they work. 0 comments __loam Reply No comments yet Contribute on Hacker News ↗
No comments yet
Contribute on Hacker News ↗