Robot System Assistent

RoSA

image

RoSA Project

RoSA (Robot System Assistant) is an intuitive multimodal interface for interacting with robots through speech, touch, and gestures. Developed through a multi-year research project, RoSA has evolved from a Wizard-of-Oz prototype into a robust control interface, evaluated in both on-site and remote industrial settings. It supports natural human–robot collaboration by simplifying robot selection, task assignment, and feedback in dynamic multi-robot environments.

image

image image
RoSA: Evaluation of Touch and Speech Input Modalities for On-Site HRI and Telerobotics
June 06, 2025
Exploring hybrid local and remote human–robot interaction through touch and speech input with RoSA 3.
image image
Multimodal Engagement Prediction in Multiperson Human–Robot Interaction
June 13, 2022
A real-time multimodal system for predicting engagement and disengagement in collaborative industrial HRI, using deep learning and rule-based approaches.
image image
Face Recognition and Tracking Framework for Human–Robot Interaction
May 30, 2022
A real-time face recognition and tracking system for intuitive HRI, integrating lightweight CNNs and ROS.
image image
RoSA: Towards Intuitive Multi-Modal and Multi-Device Human–Robot Interaction
January 28, 2022
Building on a prior Wizard-of-Oz study, this paper presents a fully autonomous RoSA system using speech, face, and gesture input...
image image
RoSA: Concept for an Intuitive Multi-Modal and Multi-Device Interaction System
October 27, 2021
Conceptual framework for RoSA — a multi-modal, multi-device human–robot interaction assistant with speech, gesture, and facial recognition.
image image
Robots and Wizards: An Investigation Into Natural Human–Robot Interaction
November 19, 2020
This study explores natural human–robot interaction using a Wizard-of-Oz approach to understand intuitive user behavior in multimodal robot control.