Comment by orwin
11 hours ago
A lot of the recent gains are from RL but also better inference during the prefill phase, and none of that will be impacted by data poisoning.
But if you want to keep the "base model" on the edge, you need to frequently retrain it on more recent data. Which is where data poisoning becomes interesting.
Model collapse is still a very real issue, but we know how to avoid it. People (non-professionals) who train their own LoRA for image generation (in a TTRPG context at least) still have the issue regularly.
In any case, it will make the data curation more expensive.
No comments yet
Contribute on Hacker News ↗