Comment by dekhn
2 years ago
As a demo once, I trained an object detector on some vector art (high quality art, made by a UX designer) that looked like various components of burgers. I also printed the art and mounted it on magnets and used a magnetic dry board; you could put components of a burger on the board, and a real-time NN would classify the various components. I did it mainly as a joke when there was a cheeseburger emoji controversy (people prefer cheese above patty, btw).
But when I was watching I realized you could probably combine this with gesture and pose detection and build a little visual language for communicating with computers. It would be wasteful and probably not very efficient, but it was still curious how much object detection enabled building things in the real world and having it input to the computer easily.
What you imagined sounds vaguely like dynamicland from Bret Victor.
https://dynamicland.org/
The dots around the paper are encoded programs, and you can use other shapes, objects, or sigils that communicate with the computer vision system.