← Back to context

Comment by cj

1 day ago

I tried this in AI studio just now with nano banana.

Results: https://imgur.com/a/9II0Aip

The white house was the original (random photo from Google). The prompt was "What paint color would look nice? Paint the house."

> (random photo from Google)

Careful with that kind of thing.

Here, it mostly poisons your test, because that exact photo probably exists in the underlying training data and the trained network will be more or less optimized on working with it. It's really the same consideration you'd want to make when testing classifiers or other ML techs 10 years ago.

Most people taking to a task like this will be using an original photo -- missing entirely from any training date, poorly framed, unevenly lit, etc -- and you need to be careful to capture as much of that as possible when trying to evaluate how a model will work in that kind of use case.

The failure and stress points for AI tools are generally kind of alien and unfamiliar because the way they operate is totally different than the way a human operates, and if you're not especially attentive to their weird failure shapes and biases when you want to test them, or you'll easily get false positives (and false negatives) that lead you to misleading conclusions.

  • Yea, the base image was the first google image result for the search term "house". So definitely in the training set.

> The prompt was "What paint color would look nice? Paint the house."

At some point, this is probably gonna result in you coming home to a painted house and a big bill, lol.

Guess they ran out of paint - notice the upper window.

  • Oops. Original link wasn't using the Pro version. Edited the comment with an updated link.