Mobile Interactions Augmented by Wearable Computing: A Design Space and Vision

Uloženo v:
Podrobná bibliografie
Název: Mobile Interactions Augmented by Wearable Computing: A Design Space and Vision
Autoři: Thomas Olsson, Kristof Van Laerhoven, Stefan Schneegass, Sven Mayer
Zdroj: International Journal of Mobile Human Computer Interaction. 8:104-114
Informace o vydavateli: IGI Global, 2016.
Rok vydání: 2016
Témata: 05 social sciences, 0501 psychology and cognitive sciences, 7. Clean energy
Popis: Wearable computing has a huge potential to shape the way we interact with mobile devices in the future. Interaction with mobile devices is still mainly limited to visual output and tactile finger-based input. Despite the visions of next-generation mobile interaction, the hand-held form factor hinders new interaction techniques becoming commonplace. In contrast, wearable devices and sensors are intended for more continuous and close-to-body use. This makes it possible to design novel wearable-augmented mobile interaction methods – both explicit and implicit. For example, the EEG signal from a wearable breast strap could be used to identify user status and change the device state accordingly (implicit) and the optical tracking with a head-mounted camera could be used to recognize gestural input (explicit). In this paper, the authors outline the design space for how the existing and envisioned wearable devices and sensors could augment mobile interaction techniques. Based on designs and discussions in a recently organized workshop on the topic as well as other related work, the authors present an overview of this design space and highlight some use cases that underline the potential therein.
Druh dokumentu: Part of book or chapter of book
Article
ISSN: 1942-3918
1942-390X
DOI: 10.4018/978-1-5225-5484-4.ch049
DOI: 10.4018/ijmhci.2016100106
Přístupové číslo: edsair.doi.dedup.....5f49fb8f6fd23e10ac006be33bbb09cc
Databáze: OpenAIRE
Popis
Abstrakt:Wearable computing has a huge potential to shape the way we interact with mobile devices in the future. Interaction with mobile devices is still mainly limited to visual output and tactile finger-based input. Despite the visions of next-generation mobile interaction, the hand-held form factor hinders new interaction techniques becoming commonplace. In contrast, wearable devices and sensors are intended for more continuous and close-to-body use. This makes it possible to design novel wearable-augmented mobile interaction methods – both explicit and implicit. For example, the EEG signal from a wearable breast strap could be used to identify user status and change the device state accordingly (implicit) and the optical tracking with a head-mounted camera could be used to recognize gestural input (explicit). In this paper, the authors outline the design space for how the existing and envisioned wearable devices and sensors could augment mobile interaction techniques. Based on designs and discussions in a recently organized workshop on the topic as well as other related work, the authors present an overview of this design space and highlight some use cases that underline the potential therein.
ISSN:19423918
1942390X
DOI:10.4018/978-1-5225-5484-4.ch049