SURF based 3D object recognition for robot hand grasping

This paper proposes a 3D object recognition method based on 3D SURF and the derivation of the robot space transformations. In a previous work, a three fingered robot hand had been developed for grasping task. The reference position of the robot hand was programmed based on predetermined values for g...

Full description

Bibliographic Details
Published in:Pertanika Journal of Science and Technology
Main Author: Remeli N.H.; Shauri R.L.A.; Yahaya F.H.; Salleh N.M.; Nasir K.; Yassin A.I.M.
Format: Article
Language:English
Published: Universiti Putra Malaysia Press 2017
Online Access:https://www.scopus.com/inward/record.uri?eid=2-s2.0-85049132223&partnerID=40&md5=86fad1f2a2e9bdcbfab22f9992d7a113
id 2-s2.0-85049132223
spelling 2-s2.0-85049132223
Remeli N.H.; Shauri R.L.A.; Yahaya F.H.; Salleh N.M.; Nasir K.; Yassin A.I.M.
SURF based 3D object recognition for robot hand grasping
2017
Pertanika Journal of Science and Technology
25
S3

https://www.scopus.com/inward/record.uri?eid=2-s2.0-85049132223&partnerID=40&md5=86fad1f2a2e9bdcbfab22f9992d7a113
This paper proposes a 3D object recognition method based on 3D SURF and the derivation of the robot space transformations. In a previous work, a three fingered robot hand had been developed for grasping task. The reference position of the robot hand was programmed based on predetermined values for grasping two different shapes of object. The work showed successful grasping but it could not generate the reference position on its own since no external sensor was used, hence it is not fully automated. Later, a 2D Speed-Up Robust Features (SURF) and 3D point cloud algorithm were applied to calculate the object’s 3D position where the result showed that the method was capable of recognising but unable to calculate the 3D position. Thus, the present study developed 3D SURF by combining recognised images based on 2D SURF and triangulation method. The identified object grasping points then are converted to robot space using the robot’s transformation equation which is derived based on dimensions between robot and camera in the workplace. The result supported the capability of the SURF algorithm for recognising the target without fail for nine random images but produced errors in the 3D position. Meanwhile, the transformation has been successful where the calculated object positions are inclined towards the directions of actual measured positions accordingly related to robot coordinates. However, maximum error of 3.90 cm was observed due to the inaccuracy of SURF detection and human error during manual measurement which can to be solved by improving the SURF algorithm in future. © 2017 Universiti Putra Malaysia Press.
Universiti Putra Malaysia Press
1287680
English
Article

author Remeli N.H.; Shauri R.L.A.; Yahaya F.H.; Salleh N.M.; Nasir K.; Yassin A.I.M.
spellingShingle Remeli N.H.; Shauri R.L.A.; Yahaya F.H.; Salleh N.M.; Nasir K.; Yassin A.I.M.
SURF based 3D object recognition for robot hand grasping
author_facet Remeli N.H.; Shauri R.L.A.; Yahaya F.H.; Salleh N.M.; Nasir K.; Yassin A.I.M.
author_sort Remeli N.H.; Shauri R.L.A.; Yahaya F.H.; Salleh N.M.; Nasir K.; Yassin A.I.M.
title SURF based 3D object recognition for robot hand grasping
title_short SURF based 3D object recognition for robot hand grasping
title_full SURF based 3D object recognition for robot hand grasping
title_fullStr SURF based 3D object recognition for robot hand grasping
title_full_unstemmed SURF based 3D object recognition for robot hand grasping
title_sort SURF based 3D object recognition for robot hand grasping
publishDate 2017
container_title Pertanika Journal of Science and Technology
container_volume 25
container_issue S3
doi_str_mv
url https://www.scopus.com/inward/record.uri?eid=2-s2.0-85049132223&partnerID=40&md5=86fad1f2a2e9bdcbfab22f9992d7a113
description This paper proposes a 3D object recognition method based on 3D SURF and the derivation of the robot space transformations. In a previous work, a three fingered robot hand had been developed for grasping task. The reference position of the robot hand was programmed based on predetermined values for grasping two different shapes of object. The work showed successful grasping but it could not generate the reference position on its own since no external sensor was used, hence it is not fully automated. Later, a 2D Speed-Up Robust Features (SURF) and 3D point cloud algorithm were applied to calculate the object’s 3D position where the result showed that the method was capable of recognising but unable to calculate the 3D position. Thus, the present study developed 3D SURF by combining recognised images based on 2D SURF and triangulation method. The identified object grasping points then are converted to robot space using the robot’s transformation equation which is derived based on dimensions between robot and camera in the workplace. The result supported the capability of the SURF algorithm for recognising the target without fail for nine random images but produced errors in the 3D position. Meanwhile, the transformation has been successful where the calculated object positions are inclined towards the directions of actual measured positions accordingly related to robot coordinates. However, maximum error of 3.90 cm was observed due to the inaccuracy of SURF detection and human error during manual measurement which can to be solved by improving the SURF algorithm in future. © 2017 Universiti Putra Malaysia Press.
publisher Universiti Putra Malaysia Press
issn 1287680
language English
format Article
accesstype
record_format scopus
collection Scopus
_version_ 1809677908549042176