Compressing deep graph convolution network with multi-staged knowledge distillation.
Given a trained deep graph convolution network (GCN), how can we effectively compress it into a compact network without significant loss of accuracy? Compressing a trained deep GCN into a compact GCN is of great importance for implementing the model to environments such as mobile or embedded systems...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Public Library of Science (PLoS)
2021-01-01
|
Series: | PLoS ONE |
Online Access: | https://doi.org/10.1371/journal.pone.0256187 |
id |
doaj-ca3c820cee7544318b24cd0850d32610 |
---|---|
record_format |
Article |
spelling |
doaj-ca3c820cee7544318b24cd0850d326102021-08-18T04:30:09ZengPublic Library of Science (PLoS)PLoS ONE1932-62032021-01-01168e025618710.1371/journal.pone.0256187Compressing deep graph convolution network with multi-staged knowledge distillation.Junghun KimJinhong JungU KangGiven a trained deep graph convolution network (GCN), how can we effectively compress it into a compact network without significant loss of accuracy? Compressing a trained deep GCN into a compact GCN is of great importance for implementing the model to environments such as mobile or embedded systems, which have limited computing resources. However, previous works for compressing deep GCNs do not consider the multi-hop aggregation of the deep GCNs, though it is the main purpose for their multiple GCN layers. In this work, we propose MustaD (Multi-staged knowledge Distillation), a novel approach for compressing deep GCNs to single-layered GCNs through multi-staged knowledge distillation (KD). MustaD distills the knowledge of 1) the aggregation from multiple GCN layers as well as 2) task prediction while preserving the multi-hop feature aggregation of deep GCNs by a single effective layer. Extensive experiments on four real-world datasets show that MustaD provides the state-of-the-art performance compared to other KD based methods. Specifically, MustaD presents up to 4.21%p improvement of accuracy compared to the second-best KD models.https://doi.org/10.1371/journal.pone.0256187 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Junghun Kim Jinhong Jung U Kang |
spellingShingle |
Junghun Kim Jinhong Jung U Kang Compressing deep graph convolution network with multi-staged knowledge distillation. PLoS ONE |
author_facet |
Junghun Kim Jinhong Jung U Kang |
author_sort |
Junghun Kim |
title |
Compressing deep graph convolution network with multi-staged knowledge distillation. |
title_short |
Compressing deep graph convolution network with multi-staged knowledge distillation. |
title_full |
Compressing deep graph convolution network with multi-staged knowledge distillation. |
title_fullStr |
Compressing deep graph convolution network with multi-staged knowledge distillation. |
title_full_unstemmed |
Compressing deep graph convolution network with multi-staged knowledge distillation. |
title_sort |
compressing deep graph convolution network with multi-staged knowledge distillation. |
publisher |
Public Library of Science (PLoS) |
series |
PLoS ONE |
issn |
1932-6203 |
publishDate |
2021-01-01 |
description |
Given a trained deep graph convolution network (GCN), how can we effectively compress it into a compact network without significant loss of accuracy? Compressing a trained deep GCN into a compact GCN is of great importance for implementing the model to environments such as mobile or embedded systems, which have limited computing resources. However, previous works for compressing deep GCNs do not consider the multi-hop aggregation of the deep GCNs, though it is the main purpose for their multiple GCN layers. In this work, we propose MustaD (Multi-staged knowledge Distillation), a novel approach for compressing deep GCNs to single-layered GCNs through multi-staged knowledge distillation (KD). MustaD distills the knowledge of 1) the aggregation from multiple GCN layers as well as 2) task prediction while preserving the multi-hop feature aggregation of deep GCNs by a single effective layer. Extensive experiments on four real-world datasets show that MustaD provides the state-of-the-art performance compared to other KD based methods. Specifically, MustaD presents up to 4.21%p improvement of accuracy compared to the second-best KD models. |
url |
https://doi.org/10.1371/journal.pone.0256187 |
work_keys_str_mv |
AT junghunkim compressingdeepgraphconvolutionnetworkwithmultistagedknowledgedistillation AT jinhongjung compressingdeepgraphconvolutionnetworkwithmultistagedknowledgedistillation AT ukang compressingdeepgraphconvolutionnetworkwithmultistagedknowledgedistillation |
_version_ |
1721203698985598976 |