Uncertain semantics, representation nuisances, and necessary invariance properties of bootstrapping agents

From Murray Wiki
Revision as of 06:16, 15 May 2016 by Murray (talk | contribs) (htdb2wiki: creating page for 2011d_cm11-icdl.html)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigationJump to search


A. Censi and R. M. Murray
IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL/EpiRob)

In the problem of bootstrapping, an agent must learn to use an unknown body, in an unknown world, starting from zero information about the world, its sensors, and its actuators. So far, this fascinating problem has not been given a proper formalization. In this paepr, we provide a possible rigorous definition of one of the key aspects of bootstrapping, namely the fact that an agent must be able to use âuninterpretedâ observations and commands. We show that this can be formalized by positing the existence of representation nuisances that act on the data, and which must be tolerated by an agent. The classes of nuisances tolerated indirectly encode the assumptions needed about the world, and therefore the agent's ability to solve smaller or larger classes of bootstrapping instances. Moreover, we argue that the behavior of an agent that claims optimality must actually be invariant to the representation nuisances, and we discuss several design principles to obtain such invariance.