Efficiently mapping high-performance early vision algorithms onto multicore embedded platforms
The combination of low-cost imaging chips and high-performance, multicore, embedded processors heralds a new era in portable vision systems. Early vision algorithms have the potential for highly data-parallel, integer execution. However, an implementation must operate within the constraints of embed...
Main Author: | |
---|---|
Published: |
Georgia Institute of Technology
2009
|
Subjects: | |
Online Access: | http://hdl.handle.net/1853/28256 |
id |
ndltd-GATECH-oai-smartech.gatech.edu-1853-28256 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-GATECH-oai-smartech.gatech.edu-1853-282562013-01-07T20:31:39ZEfficiently mapping high-performance early vision algorithms onto multicore embedded platformsApewokin, SenyoComputer visionEmbeddedMulticoreComputer visionAlgorithmsThe combination of low-cost imaging chips and high-performance, multicore, embedded processors heralds a new era in portable vision systems. Early vision algorithms have the potential for highly data-parallel, integer execution. However, an implementation must operate within the constraints of embedded systems including low clock rate, low-power operation and with limited memory. This dissertation explores new approaches to adapt novel pixel-based vision algorithms for tomorrow's multicore embedded processors. It presents : - An adaptive, multimodal background modeling technique called Multimodal Mean that achieves high accuracy and frame rate performance with limited memory and a slow-clock, energy-efficient, integer processing core. - A new workload partitioning technique to optimize the execution of early vision algorithms on multi-core systems. - A novel data transfer technique called cat-tail dma that provides globally-ordered, non-blocking data transfers on a multicore system. By using efficient data representations, Multimodal Mean provides comparable accuracy to the widely used Mixture of Gaussians (MoG) multimodal method. However, it achieves a 6.2x improvement in performance while using 18% less storage than MoG while executing on a representative embedded platform. When this algorithm is adapted to a multicore execution environment, the new workload partitioning technique demonstrates an improvement in execution times of 25% with only a 125 ms system reaction time. It also reduced the overall number of data transfers by 50%. Finally, the cat-tail buffering technique reduces the data-transfer latency between execution cores and main memory by 32.8% over the baseline technique when executing Multimodal Mean. This technique concurrently performs data transfers with code execution on individual cores, while maintaining global ordering through low-overhead scheduling to prevent collisions.Georgia Institute of Technology2009-06-08T19:33:35Z2009-06-08T19:33:35Z2009-01-09Dissertationhttp://hdl.handle.net/1853/28256 |
collection |
NDLTD |
sources |
NDLTD |
topic |
Computer vision Embedded Multicore Computer vision Algorithms |
spellingShingle |
Computer vision Embedded Multicore Computer vision Algorithms Apewokin, Senyo Efficiently mapping high-performance early vision algorithms onto multicore embedded platforms |
description |
The combination of low-cost imaging chips and high-performance, multicore, embedded processors heralds a new era in portable vision systems. Early vision algorithms have the potential for highly data-parallel, integer execution. However, an implementation must operate within the constraints of embedded systems including low clock rate, low-power operation and with limited memory. This dissertation explores new approaches to adapt novel pixel-based vision algorithms for tomorrow's multicore embedded processors. It presents :
- An adaptive, multimodal background modeling technique called Multimodal Mean that achieves high accuracy and frame rate performance with limited memory and a slow-clock, energy-efficient, integer processing core.
- A new workload partitioning technique to optimize the execution of early vision algorithms on multi-core systems.
- A novel data transfer technique called cat-tail dma that provides globally-ordered, non-blocking data transfers on a multicore system.
By using efficient data representations, Multimodal Mean provides comparable accuracy to the widely used Mixture of Gaussians (MoG) multimodal method. However, it achieves a 6.2x improvement in performance while using 18% less storage than MoG while executing on a representative embedded platform.
When this algorithm is adapted to a multicore execution environment, the new workload partitioning technique demonstrates an improvement in execution times of 25% with only a 125 ms system reaction time. It also reduced the overall number of data transfers by 50%.
Finally, the cat-tail buffering technique reduces the data-transfer latency between execution cores and main memory by 32.8% over the baseline technique when executing Multimodal Mean. This technique concurrently performs data transfers with code execution on individual cores, while maintaining global ordering through low-overhead scheduling to prevent collisions. |
author |
Apewokin, Senyo |
author_facet |
Apewokin, Senyo |
author_sort |
Apewokin, Senyo |
title |
Efficiently mapping high-performance early vision algorithms onto multicore embedded platforms |
title_short |
Efficiently mapping high-performance early vision algorithms onto multicore embedded platforms |
title_full |
Efficiently mapping high-performance early vision algorithms onto multicore embedded platforms |
title_fullStr |
Efficiently mapping high-performance early vision algorithms onto multicore embedded platforms |
title_full_unstemmed |
Efficiently mapping high-performance early vision algorithms onto multicore embedded platforms |
title_sort |
efficiently mapping high-performance early vision algorithms onto multicore embedded platforms |
publisher |
Georgia Institute of Technology |
publishDate |
2009 |
url |
http://hdl.handle.net/1853/28256 |
work_keys_str_mv |
AT apewokinsenyo efficientlymappinghighperformanceearlyvisionalgorithmsontomulticoreembeddedplatforms |
_version_ |
1716475104753352704 |