Ordinal inverse reinforcement learning applied to robot learning with small data

Autor: Colomé Figueras, Adrià, Torras, Carme
Přispěvatelé: European Commission, Universitat Politècnica de Catalunya. Departament de Matemàtiques, Institut de Robòtica i Informàtica Industrial, Universitat Politècnica de Catalunya. ROBiri - Grup de Percepció i Manipulació Robotitzada de l'IRI
Rok vydání: 2022
Předmět:
Zdroj: IEEE International Conference on Intelligent Robots and Systems 1: 2490-2496 (2022).
ISSN: 2153-0858
Popis: Trabajo presentado en el International Conference on Intelligent Robots and Systems (IROS), celebrado en Kyoto (Japón), del 23 al 27 de octubre de 2022
Over the last decade, the ability to teach actions to robots in a user-friendly way has gained relevance, and a practical way of teaching robots a new task is to use Inverse Reinforcement Learning (IRL). In IRL, an expert teacher shows the robot a desired behaviour and an agent builds a model of the reward. The agent can also infer a policy that performs in an optimal way within the limitations of the knowledge provided to it. However, most IRL approaches assume an (almost) optimal performance of the teaching agent, which might become unpractical if the teacher is not actually an expert. In addition, most IRL focus on discrete state-action spaces that limit their applicability to certain real-world problems such as within the context of direct Policy Search (PS) reinforcement learning. Therefore, in this paper we introduce Ordinal Inverse Reinforcement Learning (OrdIRL) for continuous state variables, in which the teacher can qualitatively evaluate robot performance by selecting one among the predefined performance levels (e.g. bad, medium, goodu for three tiers of performance). Once the OrdIRL has fit an ordinal distribution to the data, we propose to use Bayesian Optimization (BO) to either gain knowledge on the inferred model (exploration) or find a policy or action that maximizes the expected reward given the prior knowledge on the reward (exploitation). In the case of large-dimensional state-action spaces, we use Dimensionality Reduction (DR) techniques and perform the BO in the latent space. Experimental results on simulation and with a robot arm show how this approach allows for learning the reward function with small data.
CLOTHILDE - CLOTH manIpulation Learning from DEmonstrations (EC-H2020-741930)
Databáze: OpenAIRE