Comment by libraryofbabel
5 months ago
The papers from Anthropic on interpretability are pretty good. They look at how certain concepts are encoded within the LLM.
5 months ago
The papers from Anthropic on interpretability are pretty good. They look at how certain concepts are encoded within the LLM.
No comments yet
Contribute on Hacker News ↗