Using multi-modal 3D contours and their relations for vision and robotics

Başeski, E., Pugeault, N. , Kalkan, S., Bodenhagen, L., Piater, J. H. and Krüger, N. (2010) Using multi-modal 3D contours and their relations for vision and robotics. Journal of Visual Communication and Image Representation, 21(8), pp. 850-864. (doi: 10.1016/j.jvcir.2010.06.006)

Full text not currently available from Enlighten.

Abstract

In this work, we make use of 3D contours and relations between them (namely, coplanarity, cocolority, distance and angle) for four different applications in the area of computer vision and vision-based robotics. Our multi-modal contour representation covers both geometric and appearance information. We show the potential of reasoning with global entities in the context of visual scene analysis for driver assistance, depth prediction, robotic grasping and grasp learning. We argue that, such 3D global reasoning processes complement widely-used 2D local approaches such as bag-of-features since 3D relations are invariant under camera transformations and 3D information can be directly linked to actions. We therefore stress the necessity of including both global and local features with different spatial dimensions within a representation. We also discuss the importance of an efficient use of the uncertainty associated with the features, relations, and their applicability in a given context.

Item Type:Articles
Status:Published
Refereed:Yes
Glasgow Author(s) Enlighten ID:Pugeault, Dr Nicolas
Authors: Başeski, E., Pugeault, N., Kalkan, S., Bodenhagen, L., Piater, J. H., and Krüger, N.
College/School:College of Science and Engineering > School of Computing Science
Journal Name:Journal of Visual Communication and Image Representation
Publisher:Elsevier
ISSN:1047-3203
ISSN (Online):1095-9076
Published Online:05 July 2010

University Staff: Request a correction | Enlighten Editors: Update this record