Please use this identifier to cite or link to this item:
Title: Bayesian Cognitive Models for 3D Structure and Motion Multimodal Perception
Authors: Ferreira, João Filipe 
Orientador: Dias, Jorge Manuel Miranda
Castelo-Branco, Miguel de Sá e Sousa
Issue Date: 11-Jul-2011
Citation: FERREIRA, João Filipe - Bayesian Cognitive Models for 3D Structure and Motion Multimodal Perception [em linha]. Coimbra : [s.n], 2011. [Consult. Dia Mês Ano]. Tese de doutoramento. Disponível na WWW:<>
Serial title, monograph or event: Bayesian Cognitive Models for 3D Structure and Motion Multimodal Perception
Place of publication or event: Coimbra
Abstract: Humans use various sensory cues to extract crucial information from the environment. With a view of having robots as human companions, we are motivated towards helping to develop a knowledge representation system along the lines of what we know about us. While recent research has shown interesting results, we are still far from having concepts and algorithms that interpret space, coping with the complexity of the environment. By understanding how animals (humans) navigate and build their own spatial representation, the observed phenomena can be applied in robotics. In order to have a robust and reliable framework for navigation (i.e. in order to move within an environment, manipulate objects in it, avoid undesirable mishaps | e.g. collisions | etc.) space representation, localisation, mapping and perception are all needed. The goal of this work was to research Bayesian models to deal with fusion, multimodality, con icts, and ambiguities in perception, while simultaneously drawing inspiration upon human perceptual processes and respective behaviours. We will present a Bayesian framework for active multimodal perception of 3D structure and motion which, while not strictly neuromimetic, nds its roots in the role of the dorsal perceptual pathway of the human brain. Its composing models build upon a common egocentric spatial con guration that is naturally tting for the integration of readings from multiple sensors using a Bayesian approach. At its most basic level, these models present efficient and robust probabilistic solutions for cyclopean geometry-based stereovision and auditory perception based only on binaural cues, de ned using a consistent formalisation that allows their use as building blocks for the multimodal sensor fusion framework, both explicitly or implicitly addressing the most important challenges of sensor fusion, for vision, audition and proprioception (including vestibular sensing). Parallely, baseline research on human multimodal motion perception presented in this text provides the support for future work in new sensor models for the framework. This framework is then extended in a hierarchical fashion by incrementally implementing active perception strategies, such as active exploration based on entropy of the perceptual map that constitutes the basis of the framework and sensory saliency-based behaviours. The computational models described herewith support a real-time robotic implementation of multimodal active perception to be used in real-world applications, such as human-machine interaction or mobile robot navigation. With this work, we also hope to be able to address questions such as: Where are the limits on optimal sensory integration behaviour? What are the temporal aspects of sensory integration? How do we solve the \correspondence problem" for sensory integration? How to answer the combination versus integration debate? How to answer the switching versus weighing controversy? What are the limits of crossmodal plasticity?
Description: Tese de doutoramento em Engenharia Electrotécnica, na especialidade de Instrumentação e Controlo, apresentada à Faculdade de Ciências e Tecnologia da Universidade de Coimbra
Rights: openAccess
Appears in Collections:FCTUC Eng.Electrotécnica - Teses de Doutoramento

Files in This Item:
File Description SizeFormat
Bayesian Cognitive Models_Errata_Abbrev-1.pdf13.45 MBAdobe PDFView/Open
Show full item record

Page view(s) 50

checked on Aug 9, 2022


checked on Aug 9, 2022

Google ScholarTM


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.