Comment by pavlov

1 day ago

I’ve seen those A/B choices on Google AI Studio recently, and there wasn’t a substantial difference between the outputs. It felt more like a different random seed for the same model.

Of course it’s very possible my use case wasn’t terribly interesting so it wouldn’t reveal model differences, or that it was a different A/B test.

For me they've been very similar, except in one case where I corrected it and on one side it doubled down on being objectively wrong, and on the other side it took my feedback and started over with a new line of thinking.