This paper (as early as 1995) illustrates the idea of realizing situation-aware human-real world interaction that happens in the era of prevailing portable devices. The paper starts with pointing out three disadvantages of conventional UI when “computers are becoming increasingly portable and ubiquitous” – 1) explicity operations, 2) unaware of the real world situations, and 3) gaps between the computer world and the real world. Aiming at resolving such disadvantages, the paper proposed a concept called ‘augmented interaction’ – “a style of human-computer interaction that aims to reduce computer manipulations by using environmental information as implicit input.” This idea is further illustrated in their NAVICAM system where users, with a palmtop computer and a head-up display, are able to view superimposed messages related to specific situational context such as standing in front of a painting brings information about the artist, etc.
A comparison of HCI styles
- GUI – HCI and HRI are separated
- Virtual reality – HCI only (though it simulates the real world)
- Ubiquitous computers – HCI and HRI are more *parallel* + HCI more pervasive
- Augmented interaction – HCI, HRI, and RCI exist at the same time
- Not so sure about the statement “The main difference between (c) [Ubiquitous Computers] and (d) [Augmented Interaction] is the number of computers.”;
- What can the GUI limitations lead to?
- Explicit operations -> implicit interactions?
- Unaware of the real world situation -> context-aware?
- Gaps between the computer world and the real world -> AI? Indexing and organizing information regardless of computer-based or computerless?