Neuromorphic Eye-in-Hand Visual Servoing
Robotic vision plays a major role in factory automation to service robot applications. However, the traditional use of frame-based cameras sets a limitation on continuous visual feedback due to their low sampling rate, poor performance in low light conditions and redundant data in real-time image pr...
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9395430/ |
id |
doaj-ede817802b7d45808097c0b42b762785 |
---|---|
record_format |
Article |
spelling |
doaj-ede817802b7d45808097c0b42b7627852021-04-14T23:00:18ZengIEEEIEEE Access2169-35362021-01-019558535587010.1109/ACCESS.2021.30712619395430Neuromorphic Eye-in-Hand Visual ServoingRajkumar Muthusamy0https://orcid.org/0000-0002-5372-0154Abdulla Ayyad1https://orcid.org/0000-0002-3006-2320Mohamad Halwani2https://orcid.org/0000-0002-8478-2729Dewald Swart3Dongming Gan4https://orcid.org/0000-0001-5327-1902Lakmal Seneviratne5https://orcid.org/0000-0001-6405-8402Yahya Zweiri6https://orcid.org/0000-0003-4331-7254Khalifa University Center for Autonomous Robotic Systems (KUCARS), Khalifa University of Science and Technology, Abu Dhabi, United Arab EmiratesKhalifa University Center for Autonomous Robotic Systems (KUCARS), Khalifa University of Science and Technology, Abu Dhabi, United Arab EmiratesKhalifa University Center for Autonomous Robotic Systems (KUCARS), Khalifa University of Science and Technology, Abu Dhabi, United Arab EmiratesResearch and Development, Strata Manufacturing PJSC, Al Ain, United Arab EmiratesSchool of Engineering Technology, Purdue University, West Lafayette, IN, USAKhalifa University Center for Autonomous Robotic Systems (KUCARS), Khalifa University of Science and Technology, Abu Dhabi, United Arab EmiratesKhalifa University Center for Autonomous Robotic Systems (KUCARS), Khalifa University of Science and Technology, Abu Dhabi, United Arab EmiratesRobotic vision plays a major role in factory automation to service robot applications. However, the traditional use of frame-based cameras sets a limitation on continuous visual feedback due to their low sampling rate, poor performance in low light conditions and redundant data in real-time image processing, especially in the case of high-speed tasks. Neuromorphic event-based vision is a recent technology that gives human-like vision capabilities such as observing the dynamic changes asynchronously at a high temporal resolution (<inline-formula> <tex-math notation="LaTeX">$1~\mu s$ </tex-math></inline-formula>) with low latency and wide dynamic range. In this paper, for the first time, we present a purely event-based visual servoing method using a neuromorphic camera in an eye-in-hand configuration for the grasping pipeline of a robotic manipulator. We devise three surface layers of active events to directly process the incoming stream of events from relative motion. A purely event-based approach is used to detect corner features, localize them robustly using heatmaps and generate virtual features for tracking and grasp alignment. Based on the visual feedback, the motion of the robot is controlled to make the temporal upcoming event features converge to the desired event in Spatio-temporal space. The controller switches its operation such that it explores the workspace, reaches the target object and achieves a stable grasp. The event-based visual servoing (EBVS) method is comprehensively studied and validated experimentally using a commercial robot manipulator in an eye-in-hand configuration for both static and dynamic targets. Experimental results show superior performance of the EBVS method over frame-based vision, especially in high-speed operations and poor lighting conditions. As such, EBVS overcomes the issues of motion blur, lighting and exposure timing that exist in conventional frame-based visual servoing methods.https://ieeexplore.ieee.org/document/9395430/Neuromorphic vision sensorevent cameraevent-based visual servoingrobotic visionrobotic manipulatorneuromorphic vision-based robot control |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Rajkumar Muthusamy Abdulla Ayyad Mohamad Halwani Dewald Swart Dongming Gan Lakmal Seneviratne Yahya Zweiri |
spellingShingle |
Rajkumar Muthusamy Abdulla Ayyad Mohamad Halwani Dewald Swart Dongming Gan Lakmal Seneviratne Yahya Zweiri Neuromorphic Eye-in-Hand Visual Servoing IEEE Access Neuromorphic vision sensor event camera event-based visual servoing robotic vision robotic manipulator neuromorphic vision-based robot control |
author_facet |
Rajkumar Muthusamy Abdulla Ayyad Mohamad Halwani Dewald Swart Dongming Gan Lakmal Seneviratne Yahya Zweiri |
author_sort |
Rajkumar Muthusamy |
title |
Neuromorphic Eye-in-Hand Visual Servoing |
title_short |
Neuromorphic Eye-in-Hand Visual Servoing |
title_full |
Neuromorphic Eye-in-Hand Visual Servoing |
title_fullStr |
Neuromorphic Eye-in-Hand Visual Servoing |
title_full_unstemmed |
Neuromorphic Eye-in-Hand Visual Servoing |
title_sort |
neuromorphic eye-in-hand visual servoing |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2021-01-01 |
description |
Robotic vision plays a major role in factory automation to service robot applications. However, the traditional use of frame-based cameras sets a limitation on continuous visual feedback due to their low sampling rate, poor performance in low light conditions and redundant data in real-time image processing, especially in the case of high-speed tasks. Neuromorphic event-based vision is a recent technology that gives human-like vision capabilities such as observing the dynamic changes asynchronously at a high temporal resolution (<inline-formula> <tex-math notation="LaTeX">$1~\mu s$ </tex-math></inline-formula>) with low latency and wide dynamic range. In this paper, for the first time, we present a purely event-based visual servoing method using a neuromorphic camera in an eye-in-hand configuration for the grasping pipeline of a robotic manipulator. We devise three surface layers of active events to directly process the incoming stream of events from relative motion. A purely event-based approach is used to detect corner features, localize them robustly using heatmaps and generate virtual features for tracking and grasp alignment. Based on the visual feedback, the motion of the robot is controlled to make the temporal upcoming event features converge to the desired event in Spatio-temporal space. The controller switches its operation such that it explores the workspace, reaches the target object and achieves a stable grasp. The event-based visual servoing (EBVS) method is comprehensively studied and validated experimentally using a commercial robot manipulator in an eye-in-hand configuration for both static and dynamic targets. Experimental results show superior performance of the EBVS method over frame-based vision, especially in high-speed operations and poor lighting conditions. As such, EBVS overcomes the issues of motion blur, lighting and exposure timing that exist in conventional frame-based visual servoing methods. |
topic |
Neuromorphic vision sensor event camera event-based visual servoing robotic vision robotic manipulator neuromorphic vision-based robot control |
url |
https://ieeexplore.ieee.org/document/9395430/ |
work_keys_str_mv |
AT rajkumarmuthusamy neuromorphiceyeinhandvisualservoing AT abdullaayyad neuromorphiceyeinhandvisualservoing AT mohamadhalwani neuromorphiceyeinhandvisualservoing AT dewaldswart neuromorphiceyeinhandvisualservoing AT dongminggan neuromorphiceyeinhandvisualservoing AT lakmalseneviratne neuromorphiceyeinhandvisualservoing AT yahyazweiri neuromorphiceyeinhandvisualservoing |
_version_ |
1721527040999424000 |