Disambiguating multi-modal scene representations using perceptual grouping constraints

Pugeault, N. , Wörgötter, F. and Krüger, N. (2010) Disambiguating multi-modal scene representations using perceptual grouping constraints. PLoS ONE, 5(6), e10663. (doi: 10.1371/journal.pone.0010663) (PMID:20544006) (PMCID:PMC2882939)

[img]
Preview
Text
219327.pdf - Published Version
Available under License Creative Commons Attribution.

4MB

Abstract

In its early stages, the visual system suffers from a lot of ambiguity and noise that severely limits the performance of early vision algorithms. This article presents feedback mechanisms between early visual processes, such as perceptual grouping, stereopsis and depth reconstruction, that allow the system to reduce this ambiguity and improve early representation of visual information. In the first part, the article proposes a local perceptual grouping algorithm that — in addition to commonly used geometric information — makes use of a novel multi–modal measure between local edge/line features. The grouping information is then used to: 1) disambiguate stereopsis by enforcing that stereo matches preserve groups; and 2) correct the reconstruction error due to the image pixel sampling using a linear interpolation over the groups. The integration of mutual feedback between early vision processes is shown to reduce considerably ambiguity and noise without the need for global constraints.

Item Type:Articles
Status:Published
Refereed:Yes
Glasgow Author(s) Enlighten ID:Pugeault, Dr Nicolas
Authors: Pugeault, N., Wörgötter, F., and Krüger, N.
College/School:College of Science and Engineering > School of Computing Science
Journal Name:PLoS ONE
Publisher:Public Library of Science
ISSN:1932-6203
ISSN (Online):1932-6203
Copyright Holders:Copyright © 2010 Pugeault et al.
First Published:First published in PLoS ONE 5(6): e10663
Publisher Policy:Reproduced under a Creative Commons License

University Staff: Request a correction | Enlighten Editors: Update this record