Summary: | Our senses allow us to identify objects, materials and events in the world around us,enabling us to interact effectively with our surroundings. However, perception is inherently ambiguous, in that each set of sensory data could have resulted from an infinite number of world states. To find statistically optimal solutions the brain uses the available sensory data together with its prior knowledge or experience. In addition, when there are multiple cues available they may be: (i) combined to improve precision through noise reduction; and/or (ii) recalibrated to improve accuracy through bias reduction. This thesis investigates cue combination, learning and recalibration through a series of four studies, using the Bayesian framework to model cues and their interactions. The first study finds that haptic cues to material properties are combined with visual cues to affect estimates of object gloss. It also investigates how the binocular disparity of specular highlights affects gloss estimates. This is extended in the second study, which finds that the human visual system does not employ a full geometric model of specular highlight disparity when making shape and gloss estimates. The third study replicates and extends previous findings, that auditory and visual cues to temporal events are optimally combined in adults, by demonstrating that children also optimally combine auditory and visual cues. Both adults' and children's bimodal percepts are shown to be well predicted by a `coupling prior' model of optimal partial cue combination. The fourth study finds that the visual system can learn and invoke two context-specific priors for illumination direction, using haptic shape cues to provide calibratory feedback during training. It also demonstrates that colour can be learnt as a contextual cue. The results of all these studies are considered in the context of existing work, and ideas for future research are discussed.
|