← Back to context

Comment by twosdai

2 years ago

One of the topics I didn't see discussed in this article is how we're expected to validate the results of the output of the AI.

Really liked the announcement and I think this is a great step forward. Looking forward to use it. However I don't really see how we can verify the validity of AI responses with some statistical significance.

For example, one of the video demos shows Gemini updating a graph from some scientific literature. How do we know the data it received for the graph is accurate?

It feels like to me there is a missing prompt step not shown, which is to have a competing advisarial model be prompted to validate the results of the other model with some generated code that a human could audit.

Basically when humans work together to do the work, we review each other's work. I don't see why AIs can't do the same with a human additionally verifying it.