Recursive Tree Grammar Autoencoders
Autor: | Benjamin Paaßen, Irena Koprinska, Kalina Yacef |
---|---|
Rok vydání: | 2020 |
Předmět: |
FOS: Computer and information sciences
Computer Science - Machine Learning Artificial Intelligence Tree grammars Computer Science - Neural and Evolutionary Computing Recursive neural networks Data_CODINGANDINFORMATIONTHEORY Neural and Evolutionary Computing (cs.NE) Variational autoencoders Representation learning Software Machine Learning (cs.LG) |
DOI: | 10.48550/arxiv.2012.02097 |
Popis: | Machine learning on trees has been mostly focused on trees as input to algorithms. Much less research has investigated trees as output, which has many applications, such as molecule optimization for drug discovery, or hint generation for intelligent tutoring systems. In this work, we propose a novel autoencoder approach, called recursive tree grammar autoencoder (RTG-AE), which encodes trees via a bottom-up parser and decodes trees via a tree grammar, both learned via recursive neural networks that minimize the variational autoencoder loss. The resulting encoder and decoder can then be utilized in subsequent tasks, such as optimization and time series prediction. RTG-AEs are the first model to combine variational autoencoders, grammatical knowledge, and recursive processing. Our key message is that this unique combination of all three elements outperforms models which combine any two of the three. In particular, we perform an ablation study to show that our proposed method improves the autoencoding error, training time, and optimization score on synthetic as well as real datasets compared to four baselines. We further prove that RTG-AEs parse and generate trees in linear time and are expressive enough to handle all regular tree grammars. Comment: Submitted to the ECML/PKDD Journal Track |
Databáze: | OpenAIRE |
Externí odkaz: |