Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip

碩士 === 國立中正大學 === 電機工程研究所 === 107 === Abstract For a smart vehicle system design, driver’s attention monitoring system is essential to advanced driver assistance system (ADAS). Such technology can be achieved by using face tracking to detect distraction of driver. The computing process involves comp...

Full description

Bibliographic Details
Main Authors: Ting, YI-SIANG, 丁弈翔
Other Authors: YU, YING-HAO
Format: Others
Language:zh-TW
Published: 2018
Online Access:http://ndltd.ncl.edu.tw/handle/4c3zjh
id ndltd-TW-107CCU00442004
record_format oai_dc
spelling ndltd-TW-107CCU004420042019-05-16T01:24:52Z http://ndltd.ncl.edu.tw/handle/4c3zjh Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip 基於單晶片模糊語意表達法及全連接類神經網路實現即時駕駛視線感測系統 Ting, YI-SIANG 丁弈翔 碩士 國立中正大學 電機工程研究所 107 Abstract For a smart vehicle system design, driver’s attention monitoring system is essential to advanced driver assistance system (ADAS). Such technology can be achieved by using face tracking to detect distraction of driver. The computing process involves complicated feature extraction and pattern recognition so that design concepts of small dimension, high computing performance, and low power consumption are required in order to be implement in a vehicle. For this, this study focuses on the way to realize an efficient driving eyes tracking algorithm on a single chip. In this research, the algorithm of Semantics-based Vague Image Representation(SVIR) with a Fully Connection Neural network(FCN) are implemented on a single Field Programmable Gate Array(FPGA) chip to track driver’s eye direction in real-time with economical hardware resources usage. The reliability for eyes direction detection can be verified by a high separation from FCN. Furthermore, the proposed system also avoids the use of DSP with low logic elements and memory usages. Therefore, it is especially suitable for small embedded systems. According to the experimental results of this study, with 640*480 image resolution and 80 image frames per second, it only consumes 0.52 us to finish eyes tracking. In addition, high reliability can be found from 20% separation to the other eyes directions with 70% similarity threshold at the output of FCN. These saliences demonstrate the feasibility for advanced driver assistance systems in the future. Keywords: Human Eyes Monitoring, Semantics-based Vague Image Representation, Fully Connected Neural Network, Real-time Processing System, Advanced Driver Assistance System YU, YING-HAO 余英豪 2018 學位論文 ; thesis 65 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 國立中正大學 === 電機工程研究所 === 107 === Abstract For a smart vehicle system design, driver’s attention monitoring system is essential to advanced driver assistance system (ADAS). Such technology can be achieved by using face tracking to detect distraction of driver. The computing process involves complicated feature extraction and pattern recognition so that design concepts of small dimension, high computing performance, and low power consumption are required in order to be implement in a vehicle. For this, this study focuses on the way to realize an efficient driving eyes tracking algorithm on a single chip. In this research, the algorithm of Semantics-based Vague Image Representation(SVIR) with a Fully Connection Neural network(FCN) are implemented on a single Field Programmable Gate Array(FPGA) chip to track driver’s eye direction in real-time with economical hardware resources usage. The reliability for eyes direction detection can be verified by a high separation from FCN. Furthermore, the proposed system also avoids the use of DSP with low logic elements and memory usages. Therefore, it is especially suitable for small embedded systems. According to the experimental results of this study, with 640*480 image resolution and 80 image frames per second, it only consumes 0.52 us to finish eyes tracking. In addition, high reliability can be found from 20% separation to the other eyes directions with 70% similarity threshold at the output of FCN. These saliences demonstrate the feasibility for advanced driver assistance systems in the future. Keywords: Human Eyes Monitoring, Semantics-based Vague Image Representation, Fully Connected Neural Network, Real-time Processing System, Advanced Driver Assistance System
author2 YU, YING-HAO
author_facet YU, YING-HAO
Ting, YI-SIANG
丁弈翔
author Ting, YI-SIANG
丁弈翔
spellingShingle Ting, YI-SIANG
丁弈翔
Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip
author_sort Ting, YI-SIANG
title Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip
title_short Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip
title_full Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip
title_fullStr Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip
title_full_unstemmed Real-time Driver’s Eyes Tracking System using Semantics-based Vague Image Representation and Fully Connected Neural Network on Single Chip
title_sort real-time driver’s eyes tracking system using semantics-based vague image representation and fully connected neural network on single chip
publishDate 2018
url http://ndltd.ncl.edu.tw/handle/4c3zjh
work_keys_str_mv AT tingyisiang realtimedriverseyestrackingsystemusingsemanticsbasedvagueimagerepresentationandfullyconnectedneuralnetworkonsinglechip
AT dīngyìxiáng realtimedriverseyestrackingsystemusingsemanticsbasedvagueimagerepresentationandfullyconnectedneuralnetworkonsinglechip
AT tingyisiang jīyúdānjīngpiànmóhúyǔyìbiǎodáfǎjíquánliánjiēlèishénjīngwǎnglùshíxiànjíshíjiàshǐshìxiàngǎncèxìtǒng
AT dīngyìxiáng jīyúdānjīngpiànmóhúyǔyìbiǎodáfǎjíquánliánjiēlèishénjīngwǎnglùshíxiànjíshíjiàshǐshìxiàngǎncèxìtǒng
_version_ 1719176141188628480