Multimodality Evaluation Metrics for Human-Robot Interaction Needed

A Case Study in Immersive Telerobotics

Research output: Chapter in Book/Report/Conference proceedingConference article in proceedingsScientificpeer-review

1 Citation (Scopus)

Abstract

Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.

Original languageEnglish
Title of host publicationAdvances in Human Factors in Robots and Unmanned Systems. AHFE 2017
EditorsJessie Chen
PublisherSpringer
Pages335-347
Number of pages13
ISBN (Print)978-3-319-60383-4, 978-3-319-60384-1
DOIs
Publication statusPublished - 1 Jan 2018
MoE publication typeA4 Article in a conference publication
EventInternational Conference on Applied Human Factors and Ergonomics: International Conference on Human Factors in Robots and Unmanned Systems - Los Angeles, United States
Duration: 17 Jul 201721 Jul 2017

Publication series

NameAdvances in Intelligent Systems and Computing
PublisherSpringer
Volume595
ISSN (Print)2194-5357
ISSN (Electronic)2194-5365

Conference

ConferenceInternational Conference on Applied Human Factors and Ergonomics
Abbreviated titleAHFE 2017
CountryUnited States
CityLos Angeles
Period17/07/1721/07/17

Fingerprint

Human robot interaction
Remote control
Ergonomics
Mobile robots
Robotics
Control systems

Keywords

  • human-robot interaction
  • immersion
  • metrics
  • multimodal
  • telerobotics
  • user studies
  • wearable

Cite this

Aaltonen, I., Aromaa, S., Helin, K., & Muhammad, A. (2018). Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics. In J. Chen (Ed.), Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017 (pp. 335-347). Springer. Advances in Intelligent Systems and Computing, Vol.. 595 https://doi.org/10.1007/978-3-319-60384-1_32
Aaltonen, Iina ; Aromaa, Susanna ; Helin, Kaj ; Muhammad, Ali. / Multimodality Evaluation Metrics for Human-Robot Interaction Needed : A Case Study in Immersive Telerobotics. Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. editor / Jessie Chen. Springer, 2018. pp. 335-347 (Advances in Intelligent Systems and Computing, Vol. 595).
@inproceedings{49a7247cd80943868eabec6f8aa70091,
title = "Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics",
abstract = "Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.",
keywords = "human-robot interaction, immersion, metrics, multimodal, telerobotics, user studies, wearable",
author = "Iina Aaltonen and Susanna Aromaa and Kaj Helin and Ali Muhammad",
year = "2018",
month = "1",
day = "1",
doi = "10.1007/978-3-319-60384-1_32",
language = "English",
isbn = "978-3-319-60383-4",
series = "Advances in Intelligent Systems and Computing",
publisher = "Springer",
pages = "335--347",
editor = "Jessie Chen",
booktitle = "Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017",
address = "Germany",

}

Aaltonen, I, Aromaa, S, Helin, K & Muhammad, A 2018, Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics. in J Chen (ed.), Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. Springer, Advances in Intelligent Systems and Computing, vol. 595, pp. 335-347, International Conference on Applied Human Factors and Ergonomics, Los Angeles, United States, 17/07/17. https://doi.org/10.1007/978-3-319-60384-1_32

Multimodality Evaluation Metrics for Human-Robot Interaction Needed : A Case Study in Immersive Telerobotics. / Aaltonen, Iina; Aromaa, Susanna; Helin, Kaj; Muhammad, Ali.

Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. ed. / Jessie Chen. Springer, 2018. p. 335-347 (Advances in Intelligent Systems and Computing, Vol. 595).

Research output: Chapter in Book/Report/Conference proceedingConference article in proceedingsScientificpeer-review

TY - GEN

T1 - Multimodality Evaluation Metrics for Human-Robot Interaction Needed

T2 - A Case Study in Immersive Telerobotics

AU - Aaltonen, Iina

AU - Aromaa, Susanna

AU - Helin, Kaj

AU - Muhammad, Ali

PY - 2018/1/1

Y1 - 2018/1/1

N2 - Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.

AB - Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.

KW - human-robot interaction

KW - immersion

KW - metrics

KW - multimodal

KW - telerobotics

KW - user studies

KW - wearable

UR - http://www.scopus.com/inward/record.url?scp=85041041850&partnerID=8YFLogxK

U2 - 10.1007/978-3-319-60384-1_32

DO - 10.1007/978-3-319-60384-1_32

M3 - Conference article in proceedings

SN - 978-3-319-60383-4

SN - 978-3-319-60384-1

T3 - Advances in Intelligent Systems and Computing

SP - 335

EP - 347

BT - Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017

A2 - Chen, Jessie

PB - Springer

ER -

Aaltonen I, Aromaa S, Helin K, Muhammad A. Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics. In Chen J, editor, Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. Springer. 2018. p. 335-347. (Advances in Intelligent Systems and Computing, Vol. 595). https://doi.org/10.1007/978-3-319-60384-1_32