InDex - Robot In-hand Dexterous Manipulation
InDex - Robot In-hand Dexterous Manipulation
Disciplines
Electrical Engineering, Electronics, Information Engineering (65%); Computer Sciences (35%)
Keywords
-
Roboter,
Posebestimmung,
Manipulation,
In Der Hand,
Objekt
Humans excel when dealing with everyday objects and manipulation tasks, being able to learn new skills, and to adapt in different or complex environments. This is a basic skill for our survival as well as a key feature in our world of artefacts and human-made devices. Our expert ability to use our hands results from a lifetime of learning by both observing other skilled humans and ourselves as we discover how to handle objects first hand. Unfortunately, today`s robotic hands are still unable to achieve such a high level of dexterity in comparison to humans nor are systems entirely able to understand their own potential. In order for robots to truly operate in a human world and fulfil the expectations as intelligent assistants, they must be able to manipulate a wide variety of unknown objects by mastering their capabilities of strength, finesse and subtlety. To achieve such dexterity with robotic hands, cognitive capacity is needed to deal with uncertainties in the real world and to generalise previously learned skills to new objects and tasks. Furthermore, we assert that the complexity of programming must be greatly reduced and robot autonomy must become much more natural. The InDex project aims to understand how humans perform in-hand object manipulation and to replicate the observed skilled movements with dexterous artificial hands, merging the concepts of reinforcement and transfer learning to generalise in-hand skills for multiple objects and tasks. In addition, an abstraction and representation of previous knowledge will be fundamental for the reproducibility of learned skills to different hardware. Learning will use data across multiple modalities that will be collected, annotated and assembled into a large dataset. The data and our methods will be shared with the wider research community to allow testing against benchmarks and reproduction of results. More concretely, the core objectives are: (i) to build a multi-modal artificial perception architecture that extracts data of human manipulation of objects for progressing autonomous in-hand manipulation, i.e. allowing the robot to regrasp, reorient and finely reposition objects into their final pose for operation; (ii) the creation of a multimodal dataset of in-hand manipulation tasks; (iii) the development of an advanced object modelling and recognition system, including the characterisation of object affordances and grasping properties, in order to encapsulate both explicit information and possible implicit object usages; (iv) to autonomously learn and precisely imitate human strategies in handling tasks; and (v) to build a bridge between observation and execution, allowing deployment that is independent of the robot architecture.
- Technische Universität Wien - 100%
- Véronique Perdereau, Sorbonne Université - France
- Diego Resende Faria, Aston University
Research Output
- 1 Citations
- 1 Publications
-
2022
Title Where Does It Belong? Autonomous Object Mapping in Open-World Settings DOI 10.3389/frobt.2022.828732 Type Journal Article Author Langer E Journal Frontiers in Robotics and AI Pages 828732 Link Publication