Comment by p0w3n3d
7 months ago
If I'm not mistaken, this is like a top of the iceberg. There must be a lot of post-training - e.g. fine-tuning to make the model adhere to these rules. Just saying "you MUST not" will not make the model adhere, I'd say (according to what I have recently learnt about model fine-tuning).
No comments yet
Contribute on Hacker News ↗