Comment by pmontra
1 year ago
There are two different issues.
1. AI image generation is not the right tool for some purposes. It doesn't really know the world, it does not know history, it only understands probabilities. I would also draw weird stuff for some prompts if I was subject to those limitations.
2. The way Google is trying to adapt the wrong tool to the tasks it's not good for. No matter what they try, it's still the wrong tool. You can use a F1 car to pull a manhole cover from a road but don't expect to be happy with the result (it happened again a few hours ago, sorry for the strange example.)
No no no, don't go blaming the model here.
I guarantee that you could get the current version of Gemini without the guardrails to appropriately contextualize a prompt for historical context.
It's being directly instructed to adjust prompts with heavy handed constraints the same as Dall-E.
This isn't an instance of model limitations but an instance of engineering's lack of foresight.