Comment by numba888
3 months ago
> isn't this dangerous? isn't the efficiency given at the expense of safety and interpretability?
Final text is only a small part of model's thinking. It's produced from embeddings which probably have much more in them. Each next token depends not only on previous, but all the intermediate values for all tokens. We don't know them, they are actually important and represent inner 'thinking'. So, LLM is still a black box. The result is usually A because of B. Sort of explanation for A, but where B came from we can only guess.
No comments yet
Contribute on Hacker News ↗