Comment by twosdai
2 years ago
One of the topics I didn't see discussed in this article is how we're expected to validate the results of the output of the AI.
Really liked the announcement and I think this is a great step forward. Looking forward to use it. However I don't really see how we can verify the validity of AI responses with some statistical significance.
For example, one of the video demos shows Gemini updating a graph from some scientific literature. How do we know the data it received for the graph is accurate?
It feels like to me there is a missing prompt step not shown, which is to have a competing advisarial model be prompted to validate the results of the other model with some generated code that a human could audit.
Basically when humans work together to do the work, we review each other's work. I don't see why AIs can't do the same with a human additionally verifying it.
No comments yet
Contribute on Hacker News ↗