I asked it a few questions and it responded exactly like all the other models do. Some of the questions were difficult / very specific, and it failed in the same way all the other models failed.
Great example of this general class of reasoning failure.
“AI does badly on my test therefore it’s bad”.
The correct question to ask is, of course, what is it good at? (For bonus points, think in terms of $/task rather than simply being dominant over humans.)
I asked it a few questions and it responded exactly like all the other models do. Some of the questions were difficult / very specific, and it failed in the same way all the other models failed.
Great example of this general class of reasoning failure.
“AI does badly on my test therefore it’s bad”.
The correct question to ask is, of course, what is it good at? (For bonus points, think in terms of $/task rather than simply being dominant over humans.)
"AI does badly on my test much like other AI's did before it, therefore I don't immediately see much improvement" is a fair assumption.
3 replies →