Intuitive Collaboration with Household Robots inEverydaySett
Intuitive Collaboration with Household Robots inEverydaySett
Weave: Österreich - Belgien - Deutschland - Luxemburg - Polen - Schweiz - Slowenien - Tschechien
Disciplines
Electrical Engineering, Electronics, Information Engineering (60%); Computer Sciences (40%)
Keywords
-
Roboter,
Intuitive Schnittstelle,
Wahrnehmung,
Szenenverstehen
In recent years, there has been a rapid increase in the design and development of social service robots, which are intended for being deployed for menial tasks. Such robots need to be designed to interact with people who have little or no prior experience in interacting with a robot. Naive users need to be able to control and collaborate with these robots. Therefore, further research is needed to address the needs of nonexpert users through intuitive methods of interaction. To fill this niche, we will develop methods to better understand human-robot interaction through a collaborative robotic interface for humanoid domestic service robots, where the user can interact with the robot intuitively using multimodal cues of language, gestures and gaze. The project will go beyond the state of the art by combining vision- and gaze-based modalities with language to understand the users intentions for successful task completion. We will use a user-centric approach by conducting user studies to model how a human uses gestures and gaze naturally to collaborate with a robot. These studies will be conducted for scenarios including different tables and shelves as well as tidying up an entire room to show the generality of the approach. The scenarios will be developed both in Virtual Reality (VR) and real-world environments. The VR settings will be simplified replicas of the real-world scenarios and allow for quick prototyping and initial data collection with naive users in the design phase of the project. We will then incorporate these models into a multimodal interface for human-robot interaction. For this, we will develop a graph-based representation of visual scenes that incorporate gaze information, and from which gestural information can be extracted. The novel language module will detect the presence of ambiguous speech parts to augment them with gaze, gesture, or visual information from graph-based representations. The project aims at a completely new approach on how to interact with robots and other cyber-physical systems. The idea is that a novice user can interact with a robot via natural language with intuitive gestures and gaze, as compared to previous collaborative systems that use program-like language commands. The interface will be evaluated by conducting user studies in a real-world environment, for both sets of scenarios, to study the facilitatory and inhibitory effects of employing multiple modalities in communication. We will test the hypotheses that incorporating gesture- and gaze-based information in human-robot interaction makes the human user more comfortable and more trusting towards the robot. The WEAVE project is led by Bipin Indurkhya from University of Cracow, with Michal Vavrecka, CTU Praha, and Markus Vincze from TU Wien.
- Technische Universität Wien - 100%
- Michal Vavrecka, Czech Technical University Prague - Czechia, international project partner
- Bipin Indurkhya, Jagiellonina University - Poland, international project partner
Research Output
- 44 Citations
- 12 Publications
- 1 Policies
- 1 Datasets & models
- 1 Disseminations
- 1 Scientific Awards
-
2025
Title Unrealgensyn: a framework for generating synthetic videos of Unfrequent human events DOI 10.1007/s10055-025-01146-9 Type Journal Article Author Mulero-Pérez D Journal Virtual Reality Pages 76 Link Publication -
2025
Title Shape-Biased Texture Agnostic Representations for Improved Textureless and Metallic Object Detection and 6D Pose Estimation DOI 10.1109/wacv61041.2025.00853 Type Conference Proceeding Abstract Author Hönig P Pages 8806-8815 -
2024
Title Challenges for Monocular 6-D Object Pose Estimation in Robotics DOI 10.1109/tro.2024.3433870 Type Journal Article Author Thalhammer S Journal IEEE Transactions on Robotics Pages 4065-4084 -
2023
Title TrackAgent: 6D Object Tracking via Reinforcement Learning DOI 10.1007/978-3-031-44137-0_27 Type Book Chapter Author Röhrl K Publisher Springer Nature Pages 323-335 -
2023
Title Erkennung transparenter Objekte für die Laborautomatisierung DOI 10.1007/s00502-023-01158-w Type Journal Article Author Vincze M Journal e & i Elektrotechnik und Informationstechnik Pages 519-529 Link Publication -
2023
Title Open Challenges for Monocular Single-shot 6D Object Pose Estimation DOI 10.48550/arxiv.2302.11827 Type Preprint Author Thalhammer S -
2023
Title Challenges for Monocular 6D Object Pose Estimation in Robotics DOI 10.48550/arxiv.2307.12172 Type Preprint Author Thalhammer S -
2023
Title 3D-DAT: 3D-Dataset Annotation Toolkit for Robotic Vision DOI 10.1109/icra48891.2023.10160669 Type Conference Proceeding Abstract Author Suchi M Pages 9162-9168 -
2024
Title Human-in-the-loop error detection in an object organization task with a social robot DOI 10.3389/frobt.2024.1356827 Type Journal Article Author Frijns H Journal Frontiers in Robotics and AI Pages 1356827 Link Publication -
2023
Title Challenges of Depth Estimation for Transparent Objects DOI 10.1007/978-3-031-47969-4_22 Type Book Chapter Author Weibel J Publisher Springer Nature Pages 277-288 -
2023
Title COPE: End-to-end trainable Constant Runtime Object Pose Estimation DOI 10.1109/wacv56688.2023.00288 Type Conference Proceeding Abstract Author Patten T Pages 2859-2869 -
2023
Title Object Change Detection for Autonomous Indoor Robots in Open-World Settings DOI 10.34726/hss.2023.111500 Type Other Author Langer E Link Publication
-
2024
Title MINT initiative der Stadt Wien Type Participation in a guidance/advisory committee
-
2020
Title School workshops Type Participation in an activity, workshop or similar
-
2024
Title Best Paper Award at Austrian Symposium on AI, Robotics, and Vision 2024 Type Research prize Level of Recognition Continental/International