Software with graphical interfaces for everyday work is nowadays typically built for a particular method of interfacing; for example, VLC for Linux is built using QT and follows the classical desktop menu-bar window style, while VLC for Android looks vastly different and AFAICT shares barely more with Linux VLC than the (very comprehensive) video rendering library.

Adaption to the characteristics of the currently employed interfaces is necessary, but should be more fluid: When a user sits back from his Desktop PC where he has started a video and further uses a remote control to play/pause, the program should adapt at run time, and offer appropriate means of interaction. With the EOMA68 standard on its way to production, there might be an even more compelling use case: A device can go into suspend-to-disk plugged into a desktop mouse/keyboard/monitor interface, and can wake up on a touchscreen-only phone.

I'd like to see this usable in everyday programs; below, I've collected some more examples, and plans for how this could be implemented.

Basic definitions / terms


Use case: How things should work

Open issues

Sort Me In

Incubator status

This is a very incomplete draft.

This page is part of chrysn's public personal idea incubator; go up for its other entries, or read about the idea of having an idea incubator for more information on what this is.