Flexibility of Emulation Learning from Pioneers in Nonstationary Environments
Autor: | Yu Kono, Hiroaki Wakabayashi, Moto Shinriki, Tatsuji Takahashi |
---|---|
Rok vydání: | 2020 |
Předmět: | |
Zdroj: | Advances in Intelligent Systems and Computing ISBN: 9783030398774 JSAI |
DOI: | 10.1007/978-3-030-39878-1_9 |
Popis: | This is an extension from a selected paper from JSAI2019. Social learning is crucial in acquisition of the intelligent behaviors of humans and many kinds of animals, as it makes behavior learning far more efficient than pure trial-and-error. In imitation learning, a representative form of social learning, the agent observes specific action-state pair sequences produced by another agent (expert) and reflect them into its own action. One of its implementations in reinforcement learning is the inverse reinforcement learning. We propose another form of social learning, emulation learning, which requires much less information from another agent (pioneer). In emulation learning, the agent is given only a certain level of achievement by another agent, or a record. In this study, we implement emulation learning in the reinforcement learning setting by applying a model of satisficing action policy. We show that the emulation learning algorithm works well both in stationary and non-stationary reinforcement learning tasks, breaking the often observed trade-off like relationship between efficiency and flexibility. |
Databáze: | OpenAIRE |
Externí odkaz: |