Improving the Accuracy of Pruned Network Using Knowledge Distillation

碩士 === 國立臺灣科技大學 === 電子工程系 === 106 === The introduction of Convolutional Neural Networks (CNN) in image processing field has attracted researchers to explore the applications of CNN itself. Some network designs have been proposed to reach the state of the art capability. However, the current design o...

Full description

Bibliographic Details
Main Author: SETYA WIDYAWAN PRAKOSA
Other Authors: Jenq-Shiou Leu
Format: Others
Language:en_US
Published: 2018
Online Access:http://ndltd.ncl.edu.tw/handle/u2y997
id ndltd-TW-106NTUS5428148
record_format oai_dc
spelling ndltd-TW-106NTUS54281482019-11-28T05:22:08Z http://ndltd.ncl.edu.tw/handle/u2y997 Improving the Accuracy of Pruned Network Using Knowledge Distillation Improving the Accuracy of Pruned Network Using Knowledge Distillation SETYA WIDYAWAN PRAKOSA SETYA WIDYAWAN PRAKOSA 碩士 國立臺灣科技大學 電子工程系 106 The introduction of Convolutional Neural Networks (CNN) in image processing field has attracted researchers to explore the applications of CNN itself. Some network designs have been proposed to reach the state of the art capability. However, the current design of neural network remains an issue related to the size of the model. Thus, some researchers introduce to reduce or compress the model size. The compression technique might affect the accuracy of the compressed model compared to the original one. In addition, it may influence the performance of the new model. Furthermore, we need to exploit a new scheme to enhance the accuracy of compressed network. In this study, we explore that Knowledge Distillation (KD) can be integrated to one of pruning methodologies namely pruning filters, as the compression technique, to enhance the accuracy of pruned model. From all experimental results, we conclude that incorporating KD to create a MobileNets model can enhance the accuracy of pruned network without elongating the inference time. We measured the inference time of model trained with KD is just 0.1s longer than that of without KD. Furthermore, by reducing 26.08% of the model size, the accuracy without KD is 63.65% and by incorporating KD, we can enhance to 65.37%. By reducing the size of model using pruning filters, we can deduct the size while the original size of MobileNets is 14.4 MB and reducing 26.08% can decrease the size to 11.3 MB. We also save 0.1 s inference time by compressing the size of model. Jenq-Shiou Leu 呂政修 2018 學位論文 ; thesis 42 en_US
collection NDLTD
language en_US
format Others
sources NDLTD
description 碩士 === 國立臺灣科技大學 === 電子工程系 === 106 === The introduction of Convolutional Neural Networks (CNN) in image processing field has attracted researchers to explore the applications of CNN itself. Some network designs have been proposed to reach the state of the art capability. However, the current design of neural network remains an issue related to the size of the model. Thus, some researchers introduce to reduce or compress the model size. The compression technique might affect the accuracy of the compressed model compared to the original one. In addition, it may influence the performance of the new model. Furthermore, we need to exploit a new scheme to enhance the accuracy of compressed network. In this study, we explore that Knowledge Distillation (KD) can be integrated to one of pruning methodologies namely pruning filters, as the compression technique, to enhance the accuracy of pruned model. From all experimental results, we conclude that incorporating KD to create a MobileNets model can enhance the accuracy of pruned network without elongating the inference time. We measured the inference time of model trained with KD is just 0.1s longer than that of without KD. Furthermore, by reducing 26.08% of the model size, the accuracy without KD is 63.65% and by incorporating KD, we can enhance to 65.37%. By reducing the size of model using pruning filters, we can deduct the size while the original size of MobileNets is 14.4 MB and reducing 26.08% can decrease the size to 11.3 MB. We also save 0.1 s inference time by compressing the size of model.
author2 Jenq-Shiou Leu
author_facet Jenq-Shiou Leu
SETYA WIDYAWAN PRAKOSA
SETYA WIDYAWAN PRAKOSA
author SETYA WIDYAWAN PRAKOSA
SETYA WIDYAWAN PRAKOSA
spellingShingle SETYA WIDYAWAN PRAKOSA
SETYA WIDYAWAN PRAKOSA
Improving the Accuracy of Pruned Network Using Knowledge Distillation
author_sort SETYA WIDYAWAN PRAKOSA
title Improving the Accuracy of Pruned Network Using Knowledge Distillation
title_short Improving the Accuracy of Pruned Network Using Knowledge Distillation
title_full Improving the Accuracy of Pruned Network Using Knowledge Distillation
title_fullStr Improving the Accuracy of Pruned Network Using Knowledge Distillation
title_full_unstemmed Improving the Accuracy of Pruned Network Using Knowledge Distillation
title_sort improving the accuracy of pruned network using knowledge distillation
publishDate 2018
url http://ndltd.ncl.edu.tw/handle/u2y997
work_keys_str_mv AT setyawidyawanprakosa improvingtheaccuracyofprunednetworkusingknowledgedistillation
AT setyawidyawanprakosa improvingtheaccuracyofprunednetworkusingknowledgedistillation
_version_ 1719297037897302016