There's a difference between something like a transparent background (you can run i3/picom on a potato) and having to composite many little UI elements to render a frame.
I can think of a couple of creative ways to dramatically optimize rendering of these effects. There is probably quite some batching and reordering possible without affecting correctness.
Ceteris paribus your performance is always going to be substantially worse even with tons of fancy tricks. Those also get much harder to implement when you're building a complete UI toolkit that has to support a ton of stuff rather than just writing first-party apps/OS components.
There's a difference between something like a transparent background (you can run i3/picom on a potato) and having to composite many little UI elements to render a frame.
I can think of a couple of creative ways to dramatically optimize rendering of these effects. There is probably quite some batching and reordering possible without affecting correctness.
Ceteris paribus your performance is always going to be substantially worse even with tons of fancy tricks. Those also get much harder to implement when you're building a complete UI toolkit that has to support a ton of stuff rather than just writing first-party apps/OS components.
7 replies →
The reality distortion field is back, it seems.