Comment by kaibee
2 years ago
> An LLM could be safe the first 99 times and then randomly wipe your hard drive.
So like, has anyone ever actually done enough fuzzing to see if this or other actually bad commands ever happen in practice, or are we just going on vibes here? I suppose its possible that you give it a text description to do something bad and it does, but I'm actually curious if this is just 'llms bad' vibes.
Not intentionally, but its given me incorrect SQL that feels one step away from something incredibly dangerous