Abstract
Accelerometer-based gesture control is studied as a supplementary or an
alternative interaction modality. Gesture commands freely trainable by
the user can be used for controlling external devices with handheld
wireless sensor unit. Two user studies are presented. The first study
concerns finding gestures for controlling a design environment (Smart
Design Studio), TV, VCR, and lighting. The results indicate that
different people usually prefer different gestures for the same task,
and hence it should be possible to personalise them. The second user
study concerns evaluating the usefulness of the gesture modality
compared to other interaction modalities for controlling a design
environment. The other modalities were speech, RFID-based physical
tangible objects, laser-tracked pen, and PDA stylus. The results suggest
that gestures are a natural modality for certain tasks, and can augment
other modalities. Gesture commands were found to be natural, especially
for commands with spatial association in design environment control.
Original language | English |
---|---|
Pages (from-to) | 285 - 299 |
Number of pages | 15 |
Journal | Personal and Ubiquitous Computing |
Volume | 10 |
Issue number | 5 |
DOIs | |
Publication status | Published - 2005 |
MoE publication type | A1 Journal article-refereed |
Keywords
- gesture recognition
- gesture control
- multimodal interaction
- mobile devices
- accelerometer