Layer-Level Knowledge Distillation for Deep Neural Network Learning

Autor: Hao-Ting Li, Shih-Chieh Lin, Cheng-Yeh Chen, Chen-Kuo Chiang
Jazyk: angličtina
Rok vydání: 2019
Předmět:
Zdroj: Applied Sciences, Vol 9, Iss 10, p 1966 (2019)
Druh dokumentu: article
ISSN: 2076-3417
99244438
DOI: 10.3390/app9101966
Popis: Motivated by the recently developed distillation approaches that aim to obtain small and fast-to-execute models, in this paper a novel Layer Selectivity Learning (LSL) framework is proposed for learning deep models. We firstly use an asymmetric dual-model learning framework, called Auxiliary Structure Learning (ASL), to train a small model with the help of a larger and well-trained model. Then, the intermediate layer selection scheme, called the Layer Selectivity Procedure (LSP), is exploited to determine the corresponding intermediate layers of source and target models. The LSP is achieved by two novel matrices, the layered inter-class Gram matrix and the inter-layered Gram matrix, to evaluate the diversity and discrimination of feature maps. The experimental results, demonstrated using three publicly available datasets, present the superior performance of model training using the LSL deep model learning framework.
Databáze: Directory of Open Access Journals