Kratz and Rohs explored Around-Device Interaction (ADI) in the HoverFlow prototype where they used infra-red (IR) sensors to detect users’ gestures above the device’s input area.
In particular, they obtained data from a 2×3 array of IR sensors and maintain a time window of such data entries. The gestures were captured by detecting the changes of hand-to-device proximity within this time window and further classified by comparing them to a predefined set of gesture models.
* Both SideSight and HoverFlow don’t quite address a more common use context of mobile devices – handheld with one-/two- handed interactions. In this paper, ADI is not very clearly defined – in reality probably above/back-of the device is the only possible instances;
* Figure 3 is an example of good figures;
* To read: Theremin;
* ‘Mapping of sensor data to interface actions’ section conforms with the three models: (users’) conceptual model, sensory model (what and how sensors are employed to obtain data that supports such conceptual model), and interaction model (how the interaction design maps such data to its contents).
* For MobileHCI-like paper: it seems necessary to specify technical details.