Training neural networks to encode symbols enables combinatorial generalization

Autor: Ivan Vankov, Jeffrey S. Bowers
Jazyk: angličtina
Rok vydání: 2019
Předmět:
Zdroj: Philos Trans R Soc Lond B Biol Sci
Sygma
Explore Bristol Research
Datacite
UnpayWall
ORCID
Microsoft Academic Graph
arXiv.org e-Print Archive
European Union Open Data Portal
PubMed Central
Philosophical Transactions of the Royal Society B: Biological Sciences
Vankov, I I & Bowers, J S 2019, ' Training neural networks to encode symbols enables combinatorial generalization ', Philosophical Transactions of the Royal Society B: Biological Sciences, vol. 375, no. 1791, 20190309 . https://doi.org/10.1098/rstb.2019.0309
ISSN: 1471-2970
0962-8436
Popis: Combinatorial generalization—the ability to understand and produce novel combinations of already familiar elements—is considered to be a core capacity of the human mind and a major challenge to neural network models. A significant body of research suggests that conventional neural networks cannot solve this problem unless they are endowed with mechanisms specifically engineered for the purpose of representing symbols. In this paper, we introduce a novel way of representing symbolic structures in connectionist terms—the vectors approach to representing symbols (VARS), which allows training standard neural architectures to encode symbolic knowledge explicitly at their output layers. In two simulations, we show that neural networks not only can learn to produce VARS representations, but in doing so they achieve combinatorial generalization in their symbolic and non-symbolic output. This adds to other recent work that has shown improved combinatorial generalization under some training conditions, and raises the question of whether specific mechanisms or training routines are needed to support symbolic processing. This article is part of the theme issue ‘Towards mechanistic models of meaning composition’.
Databáze: OpenAIRE