Learning Viewpoint Invariant Face Representations from Visual Experience by Temporal Association

Marian Stewart Bartlett and Terrence J. Sejnowski

In press: H. Wechsler, P.J. Phillips, V. Bruce, S. Fogelman-Soulie, T. Huang (Eds.), Face Recognition: From Theory to Applications, NATO ASI Series F. Springer-Verlag.

Abstract

In natural visual experience, different views of an object or face tend to appear in close temporal proximity. A set of simulations is presented which demonstrate how viewpoint invariant representations of faces can be developed from visual experience by capturing the temporal relationships among the input patterns. The simulations explored the interaction of temporal smoothing of activity signals with Hebbian learning (Foldiak, 1991) in both a feed-forward system and a recurrent system. The recurrent system was a generalization of a Hopfield network with a lowpass temporal filter on all unit activities. Following training on sequences of graylevel images of faces as they changed pose, multiple views of a given face fell into the same basin of attraction, and the system acquired representations of faces that were approximately viewpoint invariant.