Improving Deep Learning by Regularized Scale-Free MSE of Representations

Autor: Xufang Luo, Yunhong Wang, Mingyang Yi
Rok vydání: 2019
Předmět:
Zdroj: Neural Information Processing ISBN: 9783030367077
ICONIP (1)
Popis: Nowadays, the research for learning representations with deep neural networks (DNNs) is attracting more and more attentions. In general, most of studies focus on designing principles for learning representations when the learning model is stochastic, which is not widely used for all kinds of machine learning tasks. In this paper, we try to seek a principle to learn efficient representations for deterministic DNNs. Considering that the last DNN layer is often a linear mapping, we utilize the generalized linear model (GLM) theory to design a principle for learning representations. First, via choosing proper link functions, we build a connection between representation learning and the optimization of the last linear layer. Next, we propose a representation learning principle, scale-free mean square error (scale-free MSE), by leveraging the MSE analysis in the GLM. Here, we find that representations with a smaller scale-free MSE can lead to a better estimation of the parameters. Then, we further propose an algorithm to learn representations with a small scale-free MSE. Here, a regularization term is employed to make the scale-free MSE get small. Finally, we test our algorithm with typical DNN models in both supervised and unsupervised learning tasks. The experiment results show that our proposed algorithm is suitable to learn efficient representations over various of experiment settings.
Databáze: OpenAIRE