We describe an approach to integrated robot control, high-level planning, and action effect learning that attempts to overcome the representational difficulties that exist between these diverse areas. Our approach combines ideas from robot vision, knowledgelevel planning, and connectionist machine learning, and focuses on the representational needs of these components.We also make use of a simple representational unit called an instantiated state transition fragment (ISTF) and a related structure called an object-action complex (OAC). The goal of this work is a general approach for inducing high-level action specifications, suitable for planning, from a robot’s interactions with the world. We present a detailed overview of our approach and show how it supports the learning of certain aspects of a high-level lepresentation from low-level world state information.
|Publication status||Published - 2008|
|Event||The 6th International Cognitive Robotics Workshop (CogRob 2008) - Patras, Greece|
Duration: 21. Jul 2008 → 22. Jul 2008
Conference number: 6
|Conference||The 6th International Cognitive Robotics Workshop (CogRob 2008)|
|Period||21/07/2008 → 22/07/2008|