Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models

碩士 === 國立臺灣科技大學 === 電機工程系 === 106 === Deep learning still encounters several issues like overfitting and oversize due to the use of a large number of layers. The huge size greatly constrains performance and portability of deep learning models in limited environments like embedded and IoT devices. In...

Full description

Bibliographic Details
Main Author: Brilian Tafjira Nugraha
Other Authors: Shun-Feng Su
Format: Others
Language:en_US
Published: 2018
Online Access:http://ndltd.ncl.edu.tw/handle/2cvk37
Description
Summary:碩士 === 國立臺灣科技大學 === 電機工程系 === 106 === Deep learning still encounters several issues like overfitting and oversize due to the use of a large number of layers. The huge size greatly constrains performance and portability of deep learning models in limited environments like embedded and IoT devices. In this study, we reported our analysis of activation and gradient output and weight in each layer of the pre-trained models of VGG-16 and custom AlexNet to measure the efficiency of its layers. The efficiencies are estimated by using our measurements and compared with the manual layer reduction to validate the most relevant method. The method for multiple layer reductions is used for validation. With this found approach, the time of one-layer reduction and re-training processes on both models can save up to 9 folds and 5 folds respectively without significant accuracy reduction.