Single-handed interaction techniques for mobile and wearable computing
MetadataShow full item record
Altmetrics Handle Statistics
Altmetrics DOI Statistics
The past decade has seen the proliferation of mobile and wearable computing devices into our everyday life. Such devices are now used throughout the day for both productivity and entertainment purposes. As a result, it is important that input techniques for these devices are efficient, effective and intuitive. Further, it is important that these techniques reflect the reality of common usage patterns. In particular, supporting single-handed usage is of paramount importance, given that in many scenarios only one hand is available. As the screen size of mobile devices are getting larger, single-handed usage becomes even more problematic. At the opposite end of the scale, using small wearable devices such as smartwatches or fitness trackers often requires two hands. This thesis is concerned with the exploration, design, and evaluation of input techniques that enable practical and effective single-handed interaction on mobile and wearable devices, which empower users to achieve more with their smart devices when only one hand is available. In particular, the thesis focuses on the practicability and actual implementation of such techniques, by using built-in or low-cost sensors that are readily available. The work first motivates the thesis topic that was encountered during the early phase of study. Then, the single-handed interaction problem is tackled with two types of device form factor, both mobile and wearable. This thesis studies the problem on three types of input modalities – mid-air gesture, hand posture, on-surface gesture and three types of interaction techniques – text input, gesture, pointing. This thesis provides several techniques, interaction methods and exemplars required to explore the single-handed interaction problem. The effectiveness and efficiency of the techniques are evaluated with rigorous studies.
Thesis, PhD Doctor of Philosophy
Description of related resourcesOpisthenar dataset UIST2019 (thesis data). Yeo, H.S., University of St Andrews, 10 Oct 2019. DOI: https://doi.org/10.17630/6ca91424-7f7f-4b7b-964b-bd2c3409aaaf
Items in the St Andrews Research Repository are protected by copyright, with all rights reserved, unless otherwise indicated.