Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection
碩士 === 國立臺灣大學 === 資訊工程學研究所 === 105 === In recent years, Deep Convolutional Neural Network (CNN) has shown an impressive performance on computer vision field. The ability of learning feature representations from large training dataset makes deep CNN outperform traditional approaches with hand-crafted...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2017
|
Online Access: | http://ndltd.ncl.edu.tw/handle/38301762610955267572 |
id |
ndltd-TW-105NTU05392136 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-105NTU053921362017-10-08T04:31:25Z http://ndltd.ncl.edu.tw/handle/38301762610955267572 Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection 結合場景及物件資訊之深度卷積神經網路物件偵測 Zong-Ying Shen 沈宗穎 碩士 國立臺灣大學 資訊工程學研究所 105 In recent years, Deep Convolutional Neural Network (CNN) has shown an impressive performance on computer vision field. The ability of learning feature representations from large training dataset makes deep CNN outperform traditional approaches with hand-crafted features on object classification and detection. However, computations for deep CNN models are time consuming due to their high complexity, which makes it hardly applicable to real world application, such as Advance Driver Assistance System (ADAS). To reduce the computation complexity, several fast object detection frameworks in the literature have been proposed, such as SSD and YOLO. Although these kind of method can run at real-time, they usually struggle with dealing of small objects due to difficulty of handling smaller input image size. In this thesis, we analyze the CNN trained on both object-centric dataset and scene- centric dataset. And, we find that scene-centric CNN has better localization ability on small objects. Based on this observation, we propose a novel object detection framework which combines the feature representations learned from object-centric and scene-centric datasets with an aim to improve the accuracy on detecting especially small objects. To validate the proposed method, we evaluate our model on MSCOCO dataset, which is the most challenging object detection dataset nowadays. The experimental results show our method can actually improve the accuracy on detection of small objects, which leads to better overall results. We also evaluate our method on PASCAL VOC 2012 and KITTI on-road datasets, and the results show that our method not only can achieve state-of-the-art accuracy on both datasets but also most importantly with real-time speed. Li-Chen Fu Pei-Yung Hsiao 傅立成 蕭培墉 2017 學位論文 ; thesis 70 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立臺灣大學 === 資訊工程學研究所 === 105 === In recent years, Deep Convolutional Neural Network (CNN) has shown an impressive performance on computer vision field. The ability of learning feature representations from large training dataset makes deep CNN outperform traditional approaches with hand-crafted features on object classification and detection. However, computations for deep CNN models are time consuming due to their high complexity, which makes it hardly applicable to real world application, such as Advance Driver Assistance System (ADAS). To reduce the computation complexity, several fast object detection frameworks in the literature have been proposed, such as SSD and YOLO. Although these kind of method can run at real-time, they usually struggle with dealing of small objects due to difficulty of handling smaller input image size.
In this thesis, we analyze the CNN trained on both object-centric dataset and scene- centric dataset. And, we find that scene-centric CNN has better localization ability on small objects. Based on this observation, we propose a novel object detection framework which combines the feature representations learned from object-centric and scene-centric datasets with an aim to improve the accuracy on detecting especially small objects.
To validate the proposed method, we evaluate our model on MSCOCO dataset, which is the most challenging object detection dataset nowadays. The experimental results show our method can actually improve the accuracy on detection of small objects, which leads to better overall results. We also evaluate our method on PASCAL VOC 2012 and KITTI on-road datasets, and the results show that our method not only can achieve state-of-the-art accuracy on both datasets but also most importantly with real-time speed.
|
author2 |
Li-Chen Fu |
author_facet |
Li-Chen Fu Zong-Ying Shen 沈宗穎 |
author |
Zong-Ying Shen 沈宗穎 |
spellingShingle |
Zong-Ying Shen 沈宗穎 Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection |
author_sort |
Zong-Ying Shen |
title |
Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection |
title_short |
Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection |
title_full |
Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection |
title_fullStr |
Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection |
title_full_unstemmed |
Deep Convolutional Neural Network with Scene-centric and Object-centric Information for Object Detection |
title_sort |
deep convolutional neural network with scene-centric and object-centric information for object detection |
publishDate |
2017 |
url |
http://ndltd.ncl.edu.tw/handle/38301762610955267572 |
work_keys_str_mv |
AT zongyingshen deepconvolutionalneuralnetworkwithscenecentricandobjectcentricinformationforobjectdetection AT chénzōngyǐng deepconvolutionalneuralnetworkwithscenecentricandobjectcentricinformationforobjectdetection AT zongyingshen jiéhéchǎngjǐngjíwùjiànzīxùnzhīshēndùjuǎnjīshénjīngwǎnglùwùjiànzhēncè AT chénzōngyǐng jiéhéchǎngjǐngjíwùjiànzīxùnzhīshēndùjuǎnjīshénjīngwǎnglùwùjiànzhēncè |
_version_ |
1718552474816610304 |