The model:

The model uses operators that model the responses of cells in MT. They are tuned to the direction of motion within their receptive fields and have an excitatory and inhibitory region. These operators essentially subtract the motion from adjacent regions of the visual field.

Image of operators:

Each region of the visual field is processed by a group of these operators that differ in their preferred direction of motion and the angle of the axis between the excitatory and inhibitory regions.

Image of model:

The maximally responding operator at each location projects to the next layer of cells. These cells are tuned to a radial pattern of input, in terms of the preferred directions of motion, from the motion subtraction operators. The maximally responding cell in this layer will have a center of the radial pattern that coincides with the translation direction of motion of the observer.

Image of model:

The model works by removing rotational components of image velocity, leaving only the image velocity due to observer translation. The preferred directions of the maximally responding motion-subtraction operator across the visual field form a radial pattern.

Image of flow field for translation and rotation through 3D cloud:

Image of preferred directions of operator responses:

The model computes heading well in the presence of rotations.

In addition, the magnitude of the motion-subtraction operator's response is proportional to the change in depth between the excitatory and inhibitory regions.  Thus, we can detect the location and magnitude of depth changes in the scene:

Image of operator response magnitudes for wall in front of plane. 

Added observer rotations do not affect this response:

Image of operator responses in the presence of rotations:

The responses at a depth edge are proportional to the difference in inverse depth between the two surfaces.

Image of graph of responses:

 

 

The work is supported by NSF grants #IBN-0196068, and #IBN-0343825.