Printer Friendly

How do we perceive a moving world?

This article provides an overview of motion processing and how factors may have an impact on performing certain tasks safely, such as driving.


Dispensing opticians [COMMUNICATION]



When thinking about 'vision,' it's easy to fall into the trap of focusing on stationary letters and lights when in reality vision is hardly ever a stationary percept. Imagine you're standing at a pedestrian crossing, or sat watching your favourite film, or even just sitting down reading this CET article; these three actions are just a handful of real-life examples that require successful visual motion perception in order to be possible. (1) To elaborate, in order to perceive moving cars we need to be able to estimate the direction, acceleration and speed of the vehicle; to watch the television we need to be able to perceive 'apparent' motion from 25 stationary images per second; (2) and to read we need to be able to appropriately process and suppress information moving across our retina at speed. (3) In everyday life the parameters required for successful motion perception largely go unnoticed (see Figure 1), but at the core of this processing there is a highly organised, cortical (neural) pathway that contributes to this dynamic visual processing capability. (4) This article will provide a summary of how we are able to perceive moving images, along with detail on how this processing occurs in the brain and how our perceptual ability can vary depending on the conditions.

Early processing

At the retina, in general terms, each individual cell is responsible for processing differences in light across the receptive field (visual field coverage) of that cell. (5) In a stationary scene this would equate to a simple process of determining the location of the light within the receptive field of that particular cell, but with moving scenes, this system requires constant updates on where the light is and which way it is moving (see Figure 2). In this way, it can be thought of as responding to changes in spatial qualities of light over time. (5)

Due to the highly retinotopic organisation of these cells in the retina (that is to say, neighbouring objects in the visual field are processed by neighbouring cells in the retina and cortex) the signals derived at this stage are able to contribute towards processing the direction of a moving object. (6) These dynamic, motion signals are thought to be processed largely by a subsection of cells that contribute to the magnocellular pathway, (7) which is one of two pathways thought to be related to classification of retinal input. The second pathway (parvocellular) is largely responsible for processing spatial properties of scenes. (8) This is not to say that the contrary parvocellular pathway has no role in the processing of motion, but it is thought that it may be more involved in processing very low velocities. (9) The magnocellular pathway begins at the level of parasol retinal ganglion cells (RGCs) and transmits signals through to the cortex slightly differently to the parvocellular pathway. These differences centre on the observation that, in general, the magnocellular signals appear to process information quicker and they appear to be less colour-opponent. This means that magnocellular cells are not responsible for processing the colour of a scene, and similarly they do not contribute much towards the fine spatial detail which would be crucial for recognising or identifying stimuli in the visual world. (5,10) Instead, these cells process the information quickly and relatively broadly. One theory explaining why this pathway might prefer to process information quickly across a broad range of retinal locations lies in the evolutionary advantage that stationary objects are less likely to be potential threats to our wellbeing than moving objects. (11) This distinction allows our attention to be drawn much quicker to moving signals than stationary ones, thereby potentially alerting us to dangerous stimuli in our environment without having to consciously identify it. This also explains why cells in the peripheral retina are excellent motion detectors while maintaining relatively poor visual acuity. To test this at home, simply move an object at the very edge of your visual field. While it is in motion you will be able to detect the movement without being able to identify the object, but if you hold it still it will become seemingly invisible.

Once the signals leave the retina, they are transferred up the retinocortical pathway through the optic nerve to the optic chiasm then through to the lateral geniculate nuclei (LGN) before being processed at the cortex. (12) At the cortex, the signals are processed in the primary visual cortex (V1/striate cortex) before being projected onto further processing in higher-order areas. These areas are largely presumed to belong to a processing pathway referred to as the 'dorsal' or 'parietal' processing stream, as most of the processing is done along the top (dorsal) side of the brain. (13) However, this is now assumed to be an over-simplification and instead it is considered more technically accurate to describe what is referred to as the 'functional specialisation' of the cortex. This simply conceptualises the idea that the visual brain is a portion of cortex made up of smaller areas that are each responsible for processing a different aspect (function) of vision. (14) For example, in a typical adult human brain, it is thought that the areas V3A, (15) V3B, (16) V5+ (17-18) (including MT and MST), V6, (19) intraparietal sulcus (IPS), (20) human ventral intraparietal area (hVIP), the cingulate sulcus (CSv), (21) and the superior temporal sulcus (STS) (22) are all motion-sensitive areas (see Figure 3). To take this at face value, it may seem redundant to have so many areas dedicated to processing the same single aspect of vision: motion. However, as discussed briefly in the introduction, motion is multifaceted and highly complicated, so these areas are able to share the processing workload between them and each be responsible for processing separate motion qualities, for example, direction, optic flow, form-from-motion. (15-16) However, some areas do seem to be more important than others forming a hierarchy. For example, there is evidence from a case study of a patient who, following damage to bilateral (both left and right) V5+, became unable to perceive motion at all. (23) Instead she reported that the world moved in stop-motion as updated frames as opposed to smooth movement. This led to several problems in her life including: struggling to cross a busy road; being startled when people approached her in public environments; and not being able to accurately judge when her cup was full of water (see Figure 4). This type of motion-processing deficit has been termed cerebral akinetopsia, and shows that although there are a number of cortical areas involved in motion processing; V5+ is one that is completely crucial for successful motion perception, that is to say, without V5+, motion is not perceivable.

Once the visual motion areas in the brain have processed the signal, the information is passed on to higher areas in order to create a complete perception. To put this into context, the signal of a tennis ball moving towards you at speed will likely result in a transmission of signals to the motor cortex to allow an appropriate movement of your hand, and you would need to use separate object-selective cortices to identify the object as a tennis ball.

Types of motion

There are several ways to describe a moving stimulus, and these descriptions contribute to independent features that are typically processed separately within the cortex. For example, the speed of a moving object is different to the acceleration, and the direction of the object is different again. (12) Two slightly more clinically relevant ways of differentiating between moving stimuli describe whether the percept affects a large portion of the visual field or whether it is restricted to a small area (global versus local), (24) and whether the motion is defined by changes in luminance (first-order) or changes in texture (second-order). (25) These descriptions of motion are considered quite low-level features but they have a huge impact on our perception of the visual world. For example, if we view a field of expanding dots, each dot will be moving in a slightly different direction but overall they will appear to be moving away from a central point (global motion). However, if we focus on just one of these expanding dots, its individual direction will be pronounced, thereby giving a perception of one individual direction (local motion) as shown in Figure 5. In this way, global motion can be thought of as the accumulation of several local motion signals.

In terms of first- and second-order motion, it is difficult to provide real-life examples as these are such low-level features of motion processing that they are usually only tested in a laboratory setting. However, it is possible to provide an example of a visual illusion that comprises first-order motion known as 'beta movement'. This phenomenon occurs when objects next to each other become illuminated one after another and create a percept that the luminance is moving through the objects, much like when the colours in Christmas tree lights appear to move along the string when they are actually just lighting up one by one (see Figure 6). Second-order motion constitutes any type of motion not defined by luminance changes. It is thought that these qualities of motion may be separated early on in the visual pathway in order to be processed by different pathways in the brain (a common theme of vision), and then integrated at a later stage of processing. (26)

Distinguishing between retinal and extra-retinal signals

A necessity for processing motion accurately is the ability to be able to appropriately and quickly discriminate between image-retina (retinal) motion and eye-head (extra-retinal) motion. (27) Retinal motion occurs when the eyes are stationary and an object moves across the visual field, thereby also moving across the retina. For example, if you focus on your stationary right hand while moving your left hand towards it, it will elicit perception of a moving object (your left hand) because the image will travel across the retina. Conversely, extra-retinal motion occurs when the eyes follow a moving object. In this instance, motion is still detected but the image is stationary on the retina. This, therefore, implies the involvement of some kind of system in the brain for comparing stimulation of the extra-orbital muscles of the eyes and the signal from the retina. To maintain the current example, you would switch your focus from your stationary right hand to the movement of your left hand over to your right hand (this is called pursuit eye movement). Now you will still perceive your hand as a moving object, despite the object itself never actually moving on the retina (see Figure 7). This is clearly a key aspect of successful perception, but an important question is: how is it possible? Several scientists have investigated this over the years, and the consensus seems to be that the brain can resolve this issue following a cortical comparison being made between information received from the retina and information sent to the muscles regarding the movement of the eyes; (28) this is called the outflow theory. In essence, the theory describes how reception of a motion signal (afferent) at the cortical (neural) level encourages the neurons to quickly compare the signal with any signals that have been sent from the brain (efferent) to move the muscles of the eye. For example, the brain can deduce that if a moving retinal signal is detected and the eyes have not moved, then the object itself must have moved. This also permits the brain to accurately resolve a moving image that is stationary on the retina. One way to test this theory at home is to close one eye and lightly prod the sclera of your other eyeball through your lower lid. (12) In this case the brain has not sent a signal to your eye muscles asking them to move, but the cortex is receiving a motion signal from the retina, so you inaccurately perceive the world as moving.

Another important aspect of motion processing is to be able to ignore movement on the retina if the image is stationary. For example, in the case of reading, the words on the page remain stationary but your eyes make very quick (saccadic) movements across the page in order to read them. If the visual cortex perceived the words as moving then reading would be very difficult. In order to investigate this, one researcher studied a subset of motion-sensitive cells in the brains of monkeys who were trained to either stare at a fixed dot on a moving background, or switch their gaze between two stationary dots. (29) In both instances the background would move across the retina of the monkey but in the condition with gaze-switching, there would be no genuine motion. They found that in the presence of an extra-retinal signal (saccadic eye movements during the gaze-switching condition), the perception of motion was suppressed and the motion-sensitive cells did not respond. In more specific terms, it is thought that when there is an extra-retinal signal, the cortex may selectively inhibit the magnocellular information before, during, and after the movement of the eyes in order to make the percept appear stationary. (30)

Can accuracy of motion perception vary?

When considering motion perception across individuals, it would be safe to assume that there will be varying levels of perceptual accuracy across each person. This can be shown empirically by differing threshold levels on measures of psychophysical ability. (31) However, it is also possible, given certain circumstances, to have intra-individual variability; that is, one person's own ability may vary from day to day.

The first example of varying perceptual accuracy can be observed in scotopic (rod-mediated) conditions. In these circumstances perception is altered in several ways including loss of sensitivity to colour vision and loss of fine detail at the fovea. (32) However, in terms of motion processing other research has found that rod-mediated vision produces a perceptual bias of moving images, such that they appear to move slower than they actually are. (33)

On average, objects are perceived to be moving approximately 25% slower than their actual speed under scotopic conditions; this is thought to be related to the attenuation of signals in detectors responsible for processing high velocities. This attenuation is thought to occur due to greater temporal averaging of rod signals relative to cone signals under scotopic conditions. (34) This could have practical implications in terms of driving at night because--providing the street is well-lit--the conditions will not truly be scotopic, but if the road is dark the driver may struggle to accurately and appropriately predict the speed of other drivers.

A similar bias is produced in low-contrast conditions, (35) often researched due to the important implications associated with the presence of fog when driving. Original research using speed-matching tasks proposed that in general, low contrast simulations led to an increase in perceived speed, (36) but more recent research utilised the knowledge that true fog tends to produce a contrast gradient of high contrast near us and lower contrast as objects are positioned further away from us. This, therefore, produces poor visibility in distant vision while maintaining relatively clear visibility in near vision (see Figure 8). Using this principle, researchers found that when visibility is clearer in the near than the distant vision, subjects in a driving simulator actually overestimated their perceived speed which led to them driving slower than the legal limit. (37) They conclude that drivers should trust their instincts in the fog and drive slower in order to stay safe on the road.


Overall, perception of a moving world is a highly complicated process involving multiple stages of processing, several processing pathways, and different cortical areas to analyse the information from the retina.

Exam questions

Under the enhanced CET rules of the GOC, MCQs for this exam appear online at Please complete online by midnight on 8 April 2017. You will be unable to submit exams after this date. Please note that when taking an exam, the MCQs may require practitioners to apply additional knowledge that has not been covered in the related CET article.

CET points will be uploaded to the GOC within 10 working days. You will then need to log into your CET portfolio by clicking on 'MyGOC' on the GOC website ( to confirm your points.


Visit, and click on the 'Related CET article' title to view the article and accompanying 'references' in full.

Course code: C-54990 Deadline: 8 April 2017

Learning objectives

* Be able to advise patients on the factors that can impact upon tasks involving motion perception such as driving (Group 1.2.4)

* Understand the latest research relating to motion perception (Group 2.5.3)

* Be able to advise patients on the factors that can impact upon tasks involving motion perception such as driving (Group 1.2.4)

Dr Samantha Strong PhD, MBPsS, AFHEA

* Dr Samantha Strong is a post-doctoral researcher based in the School of Optometry and Vision Science at the University of Bradford. She holds a BSc in Psychology from the University of York, and a PhD in Vision Science at the University of Bradford. Her research involves using neuroimaging techniques such as fMRI and TMS to investigate perception of visual stimuli in the human brain.

Caption: Figure 1 Example of a moving scene

Caption: Figure 2 Outlining the difference between localisation of stationary light (L) and direction of moving light (L) across the receptive field (RF) of a single retinal cell

Caption: Figure 3 Cortical motion areas on the lateral surface of the brain shown on an inflated and unfolded right hemisphere. Light grey represents the surface area of the brain while dark grey represents the cortical folds

Caption: Figure 4 With impaired motion perception, the brain can receive slow, stationary updates which leads to several problems in everyday life such as over-pouring drinks

Caption: Figure 5 Difference between local and global motion shown in the same example. If focusing on the small rectangle on the left, it produces a local motion signal of leftward motion. If focusing on the whole picture on the right, it produces a global motion signal of expanding motion

Caption: Figure 6 A real-life example of first-order motion (beta movement) shown in sequentially flashing Christmas lights

Caption: Figure 7 Retinal and extra-retinal signals. Both circumstances produce perception of moving image on stationary background

Caption: Figure 8 Foggy conditions produce low contrast in distant vision and higher contrast in the near vision
COPYRIGHT 2017 Ten Alps Publishing
No portion of this article can be reproduced without the express written permission from the copyright holder.
Copyright 2017 Gale, Cengage Learning. All rights reserved.

Article Details
Printer friendly Cite/link Email Feedback
Author:Strong, Samantha
Publication:Optometry Today
Date:Mar 1, 2017
Previous Article:Cream of the crop.
Next Article:Ocular electrophysiologic: a beginner's guide.

Terms of use | Privacy policy | Copyright © 2021 Farlex, Inc. | Feedback | For webmasters |