Mobile Interactions Augmented by Wearable Computing: A Design Space and Vision

Saved in:
Bibliographic Details
Title: Mobile Interactions Augmented by Wearable Computing: A Design Space and Vision
Authors: Thomas Olsson, Kristof Van Laerhoven, Stefan Schneegass, Sven Mayer
Source: International Journal of Mobile Human Computer Interaction. 8:104-114
Publisher Information: IGI Global, 2016.
Publication Year: 2016
Subject Terms: 05 social sciences, 0501 psychology and cognitive sciences, 7. Clean energy
Description: Wearable computing has a huge potential to shape the way we interact with mobile devices in the future. Interaction with mobile devices is still mainly limited to visual output and tactile finger-based input. Despite the visions of next-generation mobile interaction, the hand-held form factor hinders new interaction techniques becoming commonplace. In contrast, wearable devices and sensors are intended for more continuous and close-to-body use. This makes it possible to design novel wearable-augmented mobile interaction methods – both explicit and implicit. For example, the EEG signal from a wearable breast strap could be used to identify user status and change the device state accordingly (implicit) and the optical tracking with a head-mounted camera could be used to recognize gestural input (explicit). In this paper, the authors outline the design space for how the existing and envisioned wearable devices and sensors could augment mobile interaction techniques. Based on designs and discussions in a recently organized workshop on the topic as well as other related work, the authors present an overview of this design space and highlight some use cases that underline the potential therein.
Document Type: Part of book or chapter of book
Article
ISSN: 1942-3918
1942-390X
DOI: 10.4018/978-1-5225-5484-4.ch049
DOI: 10.4018/ijmhci.2016100106
Accession Number: edsair.doi.dedup.....5f49fb8f6fd23e10ac006be33bbb09cc
Database: OpenAIRE
Description
Abstract:Wearable computing has a huge potential to shape the way we interact with mobile devices in the future. Interaction with mobile devices is still mainly limited to visual output and tactile finger-based input. Despite the visions of next-generation mobile interaction, the hand-held form factor hinders new interaction techniques becoming commonplace. In contrast, wearable devices and sensors are intended for more continuous and close-to-body use. This makes it possible to design novel wearable-augmented mobile interaction methods – both explicit and implicit. For example, the EEG signal from a wearable breast strap could be used to identify user status and change the device state accordingly (implicit) and the optical tracking with a head-mounted camera could be used to recognize gestural input (explicit). In this paper, the authors outline the design space for how the existing and envisioned wearable devices and sensors could augment mobile interaction techniques. Based on designs and discussions in a recently organized workshop on the topic as well as other related work, the authors present an overview of this design space and highlight some use cases that underline the potential therein.
ISSN:19423918
1942390X
DOI:10.4018/978-1-5225-5484-4.ch049