Comment by agent_anuj

3 days ago

This is coming straight from my experience last week. I actually tried to test this. took 30 days of my claude code sessions, about 32k conversation turns across 21 sessions and 10 projects. classified every user message - corrections, feedback, decisions, reframes. extracted about 3200 high signal training pairs. I put a lot of emphasis on my explicit corrections where I told the AI it was wrong and what the right answer was and WHY. fine tuned qwen 4B on it with qlora. the model learned my voice perfectly - during training it would say things like 'no. fix the query. you're doing 3 joins when you only need user_id' which is exactly how I talk. but thats the problem - it learned to parrot my phrasing without understanding why I made those corrections. it memorized the what, the artifact, but completely missed the how - the reasoning process that led to the correction. Title is exactly right.