Comment by minimaxir
8 days ago
I've been keeping an eye on Qwen-Edit/Wan 2.2 shenanigans and they are interesting: however actually running those types of models is too cumbersome and in the end unclear if it's actually worth it over the $0.04/image for Nano Banana.
I was skeptical about the notion of running similar models locally as well, but the person who did this (https://old.reddit.com/r/StableDiffusion/comments/1osi1q0/wa... ) swears that they generated it locally, just letting a single 5090 crunch away for a week.
If that's true, it seems worth getting past the 'cumbersome' aspects. This tech may not put Hollywood out of business, but it's clear that the process of filmmaking won't be recognizable in 10 years if amateurs can really do this in their basements today.
Neural Viz has been putting out some extremely high quality content recently, these seem to be the closest I've seen to approaching Hollywood level:
https://www.youtube.com/watch?v=5bYA2Rv2CQ8
https://www.youtube.com/watch?v=rfTnW8pl3DE
Takes a couple mouse clicks in ComfyUI
On that subject - ComfyUI is not the future of image gen. It's an experimental rope bridge.
Adobe's conference last week points to the future of image gen. Visual tools where you mold images like clay. Hands on.
Comfy appeals to the 0.01% that like toolkits like TouchDesigner, Nannou, and ShaderToy.
Got a link handy to a video of what you're referring to from Adobe's conference? Gave it a quick google but there's a lot of content. Thanks!
2 replies →