Comment by bigstrat2003
20 hours ago
No, this is a "being stupid enough to trust an LLM" problem. They are not trustworthy, and you must not ever let them take automated actions. Anyone who does that is irresponsible and will sooner or later learn the error of their ways, as this person did.
No comments yet
Contribute on Hacker News ↗