TY - GEN
T1 - Deep neural networks with knowledge instillation
AU - Yang, Fan
AU - Liu, Ninghao
AU - Du, Mengnan
AU - Zhou, Kaixiong
AU - Ji, Shuiwang
AU - Hu, Xia
N1 - Publisher Copyright:
Copyright © 2020 by SIAM.
PY - 2020
Y1 - 2020
N2 - Deep neural network (DNN) has become an effective computational tool because of its superior performance in practice. However, the generalization of DNN still largely depends on the training data, no matter in quantity or quality. In this paper, we propose a knowledge instillation framework, named NeuKI, for feed-forward DNN, aiming to enhance learning performance with the aid of knowledge. This task is particularly challenging due to the complicated nature of knowledge and numerous variants of DNN architectures. To bridge the gap, we construct a separate knowledge-DNN faithfully encoding the instilled knowledge for joint training. The core idea is to regularize the training of target-DNN with the constructed knowledge-DNN, so that the instilled knowledge can guide the model training. The proposed NeuKI is demonstrated to be applicable to both knowledge rules and constraints, where rules are encoded by structure and constraints are handled by loss. Experiments are conducted on several real-world datasets from different domains, and the results demonstrate the effectiveness of NeuKI in improving learning performance, as well as relevant data efficiency and model interpretability.
AB - Deep neural network (DNN) has become an effective computational tool because of its superior performance in practice. However, the generalization of DNN still largely depends on the training data, no matter in quantity or quality. In this paper, we propose a knowledge instillation framework, named NeuKI, for feed-forward DNN, aiming to enhance learning performance with the aid of knowledge. This task is particularly challenging due to the complicated nature of knowledge and numerous variants of DNN architectures. To bridge the gap, we construct a separate knowledge-DNN faithfully encoding the instilled knowledge for joint training. The core idea is to regularize the training of target-DNN with the constructed knowledge-DNN, so that the instilled knowledge can guide the model training. The proposed NeuKI is demonstrated to be applicable to both knowledge rules and constraints, where rules are encoded by structure and constraints are handled by loss. Experiments are conducted on several real-world datasets from different domains, and the results demonstrate the effectiveness of NeuKI in improving learning performance, as well as relevant data efficiency and model interpretability.
UR - http://www.scopus.com/inward/record.url?scp=85089199655&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85089199655&partnerID=8YFLogxK
U2 - 10.1137/1.9781611976236.42
DO - 10.1137/1.9781611976236.42
M3 - Conference contribution
AN - SCOPUS:85089199655
T3 - Proceedings of the 2020 SIAM International Conference on Data Mining, SDM 2020
SP - 370
EP - 378
BT - Proceedings of the 2020 SIAM International Conference on Data Mining, SDM 2020
A2 - Demeniconi, Carlotta
A2 - Chawla, Nitesh
PB - Society for Industrial and Applied Mathematics Publications
T2 - 2020 SIAM International Conference on Data Mining, SDM 2020
Y2 - 7 May 2020 through 9 May 2020
ER -