Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics

    Research output: Chapter in Book/Report/Conference proceedingConference article in proceedingsScientificpeer-review

    1 Citation (Scopus)

    Abstract

    Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.

    Original languageEnglish
    Title of host publicationAdvances in Human Factors in Robots and Unmanned Systems. AHFE 2017
    EditorsJessie Chen
    PublisherSpringer
    Pages335-347
    Number of pages13
    ISBN (Print)978-3-319-60383-4, 978-3-319-60384-1
    DOIs
    Publication statusPublished - 1 Jan 2018
    MoE publication typeA4 Article in a conference publication
    EventInternational Conference on Applied Human Factors and Ergonomics: International Conference on Human Factors in Robots and Unmanned Systems - Los Angeles, United States
    Duration: 17 Jul 201721 Jul 2017

    Publication series

    SeriesAdvances in Intelligent Systems and Computing
    Volume595
    ISSN2194-5357

    Conference

    ConferenceInternational Conference on Applied Human Factors and Ergonomics
    Abbreviated titleAHFE 2017
    CountryUnited States
    CityLos Angeles
    Period17/07/1721/07/17

    Fingerprint

    Human robot interaction
    Remote control
    Ergonomics
    Mobile robots
    Robotics
    Control systems

    Keywords

    • human-robot interaction
    • immersion
    • metrics
    • multimodal
    • telerobotics
    • user studies
    • wearable

    Cite this

    Aaltonen, I., Aromaa, S., Helin, K., & Muhammad, A. (2018). Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics. In J. Chen (Ed.), Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017 (pp. 335-347). Springer. Advances in Intelligent Systems and Computing, Vol.. 595 https://doi.org/10.1007/978-3-319-60384-1_32
    Aaltonen, Iina ; Aromaa, Susanna ; Helin, Kaj ; Muhammad, Ali. / Multimodality Evaluation Metrics for Human-Robot Interaction Needed : A Case Study in Immersive Telerobotics. Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. editor / Jessie Chen. Springer, 2018. pp. 335-347 (Advances in Intelligent Systems and Computing, Vol. 595).
    @inproceedings{49a7247cd80943868eabec6f8aa70091,
    title = "Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics",
    abstract = "Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.",
    keywords = "human-robot interaction, immersion, metrics, multimodal, telerobotics, user studies, wearable",
    author = "Iina Aaltonen and Susanna Aromaa and Kaj Helin and Ali Muhammad",
    year = "2018",
    month = "1",
    day = "1",
    doi = "10.1007/978-3-319-60384-1_32",
    language = "English",
    isbn = "978-3-319-60383-4",
    series = "Advances in Intelligent Systems and Computing",
    publisher = "Springer",
    pages = "335--347",
    editor = "Jessie Chen",
    booktitle = "Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017",
    address = "Germany",

    }

    Aaltonen, I, Aromaa, S, Helin, K & Muhammad, A 2018, Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics. in J Chen (ed.), Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. Springer, Advances in Intelligent Systems and Computing, vol. 595, pp. 335-347, International Conference on Applied Human Factors and Ergonomics, Los Angeles, United States, 17/07/17. https://doi.org/10.1007/978-3-319-60384-1_32

    Multimodality Evaluation Metrics for Human-Robot Interaction Needed : A Case Study in Immersive Telerobotics. / Aaltonen, Iina; Aromaa, Susanna; Helin, Kaj; Muhammad, Ali.

    Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. ed. / Jessie Chen. Springer, 2018. p. 335-347 (Advances in Intelligent Systems and Computing, Vol. 595).

    Research output: Chapter in Book/Report/Conference proceedingConference article in proceedingsScientificpeer-review

    TY - GEN

    T1 - Multimodality Evaluation Metrics for Human-Robot Interaction Needed

    T2 - A Case Study in Immersive Telerobotics

    AU - Aaltonen, Iina

    AU - Aromaa, Susanna

    AU - Helin, Kaj

    AU - Muhammad, Ali

    PY - 2018/1/1

    Y1 - 2018/1/1

    N2 - Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.

    AB - Multimodal, wearable technologies have the potential to enable a completely immersive teleoperation experience, which can be beneficial for a number of teleoperated robotic applications. To gain the full benefit of these technologies, understanding the user perspective of human-robot interaction (HRI) is of special relevance for highly advanced telerobotic systems in the future. In telerobotics research, however, the complex nature of multimodal interaction has not attracted much attention. We studied HRI with a wearable multimodal control system used for teleoperating a mobile robot, and recognized a need for evaluation metrics for multimodality. In the case study, questionnaires, interviews, observations and video analysis were used to evaluate usability, ergonomics, immersion, and the nature of multimodal interaction. Although the technical setup was challenging, our findings provide insights to the design and evaluation of user interaction of future immersive teleoperation systems. We propose new HRI evaluation metrics: Type of multimodal interaction and Wearability.

    KW - human-robot interaction

    KW - immersion

    KW - metrics

    KW - multimodal

    KW - telerobotics

    KW - user studies

    KW - wearable

    UR - http://www.scopus.com/inward/record.url?scp=85041041850&partnerID=8YFLogxK

    U2 - 10.1007/978-3-319-60384-1_32

    DO - 10.1007/978-3-319-60384-1_32

    M3 - Conference article in proceedings

    SN - 978-3-319-60383-4

    SN - 978-3-319-60384-1

    T3 - Advances in Intelligent Systems and Computing

    SP - 335

    EP - 347

    BT - Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017

    A2 - Chen, Jessie

    PB - Springer

    ER -

    Aaltonen I, Aromaa S, Helin K, Muhammad A. Multimodality Evaluation Metrics for Human-Robot Interaction Needed: A Case Study in Immersive Telerobotics. In Chen J, editor, Advances in Human Factors in Robots and Unmanned Systems. AHFE 2017. Springer. 2018. p. 335-347. (Advances in Intelligent Systems and Computing, Vol. 595). https://doi.org/10.1007/978-3-319-60384-1_32