A Temporal Stability Approach to Position and Attention-Shift-Invariant Recognition

Abstract

Incorporation of visual-related self-action signals can help neural networks learn invariance. We describe a method that can produce a network with invariance to changes in visual input caused by eye movements and covert attention shifts. Training of the network is controlled by signals associated with eye movements and covert attention shifting. A temporal perceptual stability constraint is used to drive the output of the network toward remaining constant across temporal sequences of saccadicmotions and covert attention shifts. We use a four-layer neural network model to perform the position-invariant extraction of local features and temporal integration of invariant presentations of local features in a bottom-up structure. We present results on both simulated data and real images to demonstrate that our network can acquire both position and attention shift invariance.

Cite

Text

Li and Clark. "A Temporal Stability Approach to Position and Attention-Shift-Invariant Recognition." Neural Computation, 2004. doi:10.1162/0899766041941907

Markdown

[Li and Clark. "A Temporal Stability Approach to Position and Attention-Shift-Invariant Recognition." Neural Computation, 2004.](https://mlanthology.org/neco/2004/li2004neco-temporal/) doi:10.1162/0899766041941907

BibTeX

@article{li2004neco-temporal,
  title     = {{A Temporal Stability Approach to Position and Attention-Shift-Invariant Recognition}},
  author    = {Li, Muhua and Clark, James J.},
  journal   = {Neural Computation},
  year      = {2004},
  pages     = {2293-2321},
  doi       = {10.1162/0899766041941907},
  volume    = {16},
  url       = {https://mlanthology.org/neco/2004/li2004neco-temporal/}
}