[lingtalks] Tim Marks Talk, Monday Feb. 25 at 12pm

Steven Ford sford at cogsci.ucsd.edu
Fri Feb 22 13:18:47 PST 2008


The UCSD Department of Cognitive Science is pleased to announce a talk by

Tim Marks Ph.D.

Department of Computer Science and Engineering
University of California, San Diego

Monday, February 25, 2008 at 12pm
Cognitive Science Building, room 003


"Probabilistic Models of Visual Processing"

If the brain has developed optimal solutions to the perceptual problems it 
faces, then it may be implementing elegant probabilistic solutions to the 
problems of dealing with an uncertain world. Studying human perception can 
inform new systems for machine perception, and developing new systems for 
machine perception improves our understanding of human and animal 
perception. My research takes advantage of this synergy between human and 
machine perception by developing probabilistic models and deriving optimal 
inference algorithms on those models to enable machines to perform a 
variety of visual processing tasks.

I will first discuss a generative model and probabilistic inference 
algorithm, called G-flow, for tracking a human face (or other deformable 
object) in 3D from single-camera video. Two standard computer vision 
algorithms, optical flow and template matching, emerge as special limiting 
cases of optimal inference under G-flow. This elucidates the conditions 
under which each of these existing methods is optimal and suggests a 
broader family of tracking algorithms that includes an entire continuum 
between these two extremes. Then I will discuss diffusion networks, a 
stochastic version of continuous time, continuous state recurrent neural 
networks. I will present the surprising result that a particular class of 
linear diffusion networks is equivalent to factor analysis, and demonstrate 
that this neurally plausible architecture can be trained, using contrastive 
divergence (Hinton 2002), to learn the type of 3D deformable models used by 
G-flow for face tracking. The same mathematical technique used in G-flow 
can be applied to simultaneous localization and mapping (SLAM) in mobile 
robots. I will present a new algorithm, called Gamma-SLAM, that uses stereo 
vision for SLAM in off-road outdoor environments by representing the world 
using variance grid maps. Next, I will present two probabilistic models of 
visual processing and compare them directly with human performance: NIMBLE, 
a Bayesian model of saccade-based visual memory; and SUN, a Bayesian 
framework for saliency using natural statistics. NIMBLE achieves human 
levels of performance in a face recognition task using a fixation-based 
memory model, and SUN derives a state-of-the-art saliency algorithm from 
the simple assumption that a goal of the visual system is to find potential 
targets that are important to survival. I will conclude with a discussion 
of my plans for future research in each of these areas.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://pidgin.ucsd.edu/pipermail/lingtalks/attachments/20080222/34269969/attachment.htm 


More information about the Lingtalks mailing list