Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems

碩士 === 國立臺灣大學 === 資訊工程學研究所 === 107 === Deep Neural Networks (DNNs) based learning methods have brought revolutionary advances in computer vision and machine learning. However, training a DNN model often requires very intensive computational resources. For edge incremental learning, more energy effic...

Full description

Bibliographic Details
Main Authors: Zi-Xuan Huang, 黃梓軒
Other Authors: Wei-Chung Hsu
Format: Others
Language:zh-TW
Published: 2019
Online Access:http://ndltd.ncl.edu.tw/handle/7da878
id ndltd-TW-107NTU05392136
record_format oai_dc
spelling ndltd-TW-107NTU053921362019-11-21T05:34:27Z http://ndltd.ncl.edu.tw/handle/7da878 Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems 類神經網路在異質系統架構訓練的高效能動態裝置配置方法 Zi-Xuan Huang 黃梓軒 碩士 國立臺灣大學 資訊工程學研究所 107 Deep Neural Networks (DNNs) based learning methods have brought revolutionary advances in computer vision and machine learning. However, training a DNN model often requires very intensive computational resources. For edge incremental learning, more energy efficient learning solutions are called for. Heterogeneous computing is more power efficient, and has been increasingly popular for embedded platforms. Therefore, how to deploy training models on heterogeneous platforms to support edge learning is a critical issue. Due to the increasing size of DNNs, it is rather difficult to determine how to dispatch a large number of operations to proper devices. One state-of-art approach uses reinforcement learning to address this device placement issue, but is too costly to apply in an embedded setting. In this paper, our approach leverages the information available from the computational graph of the model, and the dynamic profiles of run time and communication time of each device, to more efficiently deploy operations on heterogeneous systems. We use Critical Earliest Finish Time (CEFT) algorithm together with the Partitioned Boolean Quadratic Assignment Problem (PBQP) solver to find a cost-effective placement, and dynamically adjust assignments during the training process, which makes our method more adaptive and effective for different computational environments. On AlexNet, VGG, Inception, ResNet, RNNLM and other well-known models, our approach significantly outperforms traditional algorithms and reinforcement learning based methods. Wei-Chung Hsu 徐慰中 2019 學位論文 ; thesis 27 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 國立臺灣大學 === 資訊工程學研究所 === 107 === Deep Neural Networks (DNNs) based learning methods have brought revolutionary advances in computer vision and machine learning. However, training a DNN model often requires very intensive computational resources. For edge incremental learning, more energy efficient learning solutions are called for. Heterogeneous computing is more power efficient, and has been increasingly popular for embedded platforms. Therefore, how to deploy training models on heterogeneous platforms to support edge learning is a critical issue. Due to the increasing size of DNNs, it is rather difficult to determine how to dispatch a large number of operations to proper devices. One state-of-art approach uses reinforcement learning to address this device placement issue, but is too costly to apply in an embedded setting. In this paper, our approach leverages the information available from the computational graph of the model, and the dynamic profiles of run time and communication time of each device, to more efficiently deploy operations on heterogeneous systems. We use Critical Earliest Finish Time (CEFT) algorithm together with the Partitioned Boolean Quadratic Assignment Problem (PBQP) solver to find a cost-effective placement, and dynamically adjust assignments during the training process, which makes our method more adaptive and effective for different computational environments. On AlexNet, VGG, Inception, ResNet, RNNLM and other well-known models, our approach significantly outperforms traditional algorithms and reinforcement learning based methods.
author2 Wei-Chung Hsu
author_facet Wei-Chung Hsu
Zi-Xuan Huang
黃梓軒
author Zi-Xuan Huang
黃梓軒
spellingShingle Zi-Xuan Huang
黃梓軒
Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems
author_sort Zi-Xuan Huang
title Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems
title_short Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems
title_full Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems
title_fullStr Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems
title_full_unstemmed Efficient Dynamic Device Placement for Deep Neural NetworkTraining on Heterogeneous Systems
title_sort efficient dynamic device placement for deep neural networktraining on heterogeneous systems
publishDate 2019
url http://ndltd.ncl.edu.tw/handle/7da878
work_keys_str_mv AT zixuanhuang efficientdynamicdeviceplacementfordeepneuralnetworktrainingonheterogeneoussystems
AT huángzǐxuān efficientdynamicdeviceplacementfordeepneuralnetworktrainingonheterogeneoussystems
AT zixuanhuang lèishénjīngwǎnglùzàiyìzhìxìtǒngjiàgòuxùnliàndegāoxiàonéngdòngtàizhuāngzhìpèizhìfāngfǎ
AT huángzǐxuān lèishénjīngwǎnglùzàiyìzhìxìtǒngjiàgòuxùnliàndegāoxiàonéngdòngtàizhuāngzhìpèizhìfāngfǎ
_version_ 1719294492867035136