Human Pose Estimation Using Depth Map and Particle Swarm Optimization
碩士 === 國立臺灣大學 === 電信工程學研究所 === 100 === In this thesis, we propose a human pose estimation algorithm and implement the algorithm on CUDA platform. The proposed algorithm needs only single-view depth image as input, unlike some former works which take color images or multi-view images. The proposed al...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2012
|
Online Access: | http://ndltd.ncl.edu.tw/handle/57974942021269176114 |
id |
ndltd-TW-100NTU05435083 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-100NTU054350832015-10-13T21:50:18Z http://ndltd.ncl.edu.tw/handle/57974942021269176114 Human Pose Estimation Using Depth Map and Particle Swarm Optimization 利用深度圖與粒子群優化演算法之人體動作偵測 Chih-Chun Yang 楊智鈞 碩士 國立臺灣大學 電信工程學研究所 100 In this thesis, we propose a human pose estimation algorithm and implement the algorithm on CUDA platform. The proposed algorithm needs only single-view depth image as input, unlike some former works which take color images or multi-view images. The proposed algorithm contains the following features: first, a 32 degree-of-free model composed of two elliptic cylinder and nine ellipsoids is adopted to formulate an optimization problem. Second, a modified particle swarm optimization (PSO) scheme is applied to solve the optimization problem. And this highly parallel algorithm is suitable to be implemented on CUDA platform to achieve real-time performance. We use the Microsoft Kinect as depth sensor and use the NVIDIA GTS450 as computing device. The experimental result shows that the proposed algorithm is robust enough to overcome the self-occlusion which is the common difficulty in this area. And with the aid of this GPU, this algorithm can work in real-time (12-33 fps). Shyh-Kang Jeng 鄭士康 2012 學位論文 ; thesis 38 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立臺灣大學 === 電信工程學研究所 === 100 === In this thesis, we propose a human pose estimation algorithm and implement the algorithm on CUDA platform. The proposed algorithm needs only single-view depth image as input, unlike some former works which take color images or multi-view images. The proposed algorithm contains the following features: first, a 32 degree-of-free model composed of two elliptic cylinder and nine ellipsoids is adopted to formulate an optimization problem. Second, a modified particle swarm optimization (PSO) scheme is applied to solve the optimization problem. And this highly parallel algorithm is suitable to be implemented on CUDA platform to achieve real-time performance. We use the Microsoft Kinect as depth sensor and use the NVIDIA GTS450 as computing device. The experimental result shows that the proposed algorithm is robust enough to overcome the self-occlusion which is the common difficulty in this area. And with the aid of this GPU, this algorithm can work in real-time (12-33 fps).
|
author2 |
Shyh-Kang Jeng |
author_facet |
Shyh-Kang Jeng Chih-Chun Yang 楊智鈞 |
author |
Chih-Chun Yang 楊智鈞 |
spellingShingle |
Chih-Chun Yang 楊智鈞 Human Pose Estimation Using Depth Map and Particle Swarm Optimization |
author_sort |
Chih-Chun Yang |
title |
Human Pose Estimation Using Depth Map and Particle Swarm Optimization |
title_short |
Human Pose Estimation Using Depth Map and Particle Swarm Optimization |
title_full |
Human Pose Estimation Using Depth Map and Particle Swarm Optimization |
title_fullStr |
Human Pose Estimation Using Depth Map and Particle Swarm Optimization |
title_full_unstemmed |
Human Pose Estimation Using Depth Map and Particle Swarm Optimization |
title_sort |
human pose estimation using depth map and particle swarm optimization |
publishDate |
2012 |
url |
http://ndltd.ncl.edu.tw/handle/57974942021269176114 |
work_keys_str_mv |
AT chihchunyang humanposeestimationusingdepthmapandparticleswarmoptimization AT yángzhìjūn humanposeestimationusingdepthmapandparticleswarmoptimization AT chihchunyang lìyòngshēndùtúyǔlìziqúnyōuhuàyǎnsuànfǎzhīréntǐdòngzuòzhēncè AT yángzhìjūn lìyòngshēndùtúyǔlìziqúnyōuhuàyǎnsuànfǎzhīréntǐdòngzuòzhēncè |
_version_ |
1718068962605924352 |