Summary: Researchers have discovered a novel neural mechanism involved in casual inference that helps the brain detect objects in motion while we are moving.
Source: University of Rochester:
Imagine you’re sitting on a train. You look out the window and see another train on an adjacent track that appears to be moving. But, has your train stopped while the other train is moving, or are you moving while the other train is stopped?
The same sensory experience — viewing a train — can yield two very different perceptions, leading you to feel either a sensation of yourself in motion or a sensation of being stationary while an object moves around you.
Human brains are constantly faced with such ambiguous sensory inputs. In order to resolve the ambiguity and correctly perceive the world, our brains employ a process known as causal inference.
Causal inference is a key to learning, reasoning, and decision making, but researchers currently know little about the neurons involved in the process.
In a new paper published in the journal: eLife:researchers at the University of Rochester, including Greg DeAngelis, the George Eastman Professor of Brain and Cognitive Sciences, and his colleagues at Sungkyunkwan University and New York University, describe a novel neural mechanism involved in causal inference that helps the brain detect object motion during self-motion.
The research offers new insights into how the brain interprets sensory information and may have applications in designing artificial intelligence devices and developing treatments and therapies to treat brain disorders.
“While much has been learned previously about how the brain processes visual motion, most laboratory studies of neurons have ignored the complexities introduced by self-motion,” DeAngelis says. “Under natural conditions, identifying how objects move in the world is much more challenging for the brain.”
Now imagine a still, crouching lion waiting to spot prey; it is easy for the lion to spot a moving gazelle. Just like the still lion, when an observer is stationary, it is easy for her to detect when objects move in the world, because motion in the world directly maps to motion on the retina.
However, when the observer is also moving, her eyes are taking in motion everywhere on her retina as she moves relative to objects in the scene.
This causes a complex pattern of motion that makes it more difficult for the brain to detect when an object is moving in the world and when it is stationary; in this case, the brain has to distinguish between image motion that results from the observer itself versus image motion of other objects around the self.
The researchers discovered a type of neuron in the brain that has a particular combination of response properties, which makes the neuron well-suited to contribute to the task of distinguishing between self-motion and the motion of other objects.
“Although the brain probably uses multiple tricks to solve this problem, this new mechanism has the advantage that it can be performed in parallel at each local region of the visual field, and thus may be faster to implement than more global processes,” DeAngelis says. . “This mechanism might also be applicable to autonomous vehicles, which also need to rapidly detect moving objects.”
Unraveling a complicated circuit of neurons:
Causal inference involves a complicated circuit of neurons and other sensory mechanisms that are not widely understood, DeAngelis says, because “sensory perception works so well most of the time, so we take for granted how difficult of a computational problem it is.”
In actuality, sensory signals are noisy and incomplete. Additionally, there are many possible events that could happen in the world that would produce similar patterns of sensory input.
Consider a spot of light that moves across the retina of the eye. The same visual input could be the result of a variety of situations: it could be caused by an object that moves in the world while the viewer remains stationary, such as a person standing still at a window and observing a moving ambulance with a flashing light; it could be caused by a moving observer viewing a stationary object, such as a runner noticing a lamppost from a distance; or it could be caused by many different combinations of object motion, self-motion, and depth.
The brain has a difficult problem to solve: it must infer what most likely caused the specific pattern of sensory signals that it received. It can then draw conclusions about the situation and plan appropriate actions in response.
Building on these latest results and using data science, lab experiments, computer models, and cognitive theory, DeAngelis, Haefner, and their colleagues will continue working to pinpoint single neurons and groups of neurons that are involved in the process.
Their goal is to identify how the brain generates a consistent view of reality through interactions between the parts of the brain that process sensory stimuli and the parts of the brain that make decisions and plan actions.
Developing therapies and artificial intelligence:
Recognizing how the brain uses causal inference to separate self-motion from object motion may help in designing artificial intelligence and autopilot devices.
“Understanding how the brain infers self-motion and object motion might provide inspiration for improving existing algorithms for autopilot devices on planes and self-driving cars,” Haefner says. For example, a plane’s circuitry must take into account the plane’s self-motion in the air while also avoiding other moving planes appearing around it.
The research may additionally have important applications in developing treatments and therapies for neural disorders such as autism and schizophrenia, conditions in which casual inference is thought to be impaired.
“While the project is basic science focused on understanding the fundamental mechanisms of causal inference, this knowledge should eventually be applicable to the treatment of these disorders,” DeAngelis says.
About this visual neuroscience research news:
Author: Lindsey Valich:
Source: University of Rochester:
Contact: Lindsey Valich – University of Rochester:
Image: The image is in the public domain:
Original Research: Open access.
“A neural mechanism for detecting object motion during self-motion” by HyungGoo R Kim et al. eLife:
A neural mechanism for detecting object motion during self-motion:
Detection of objects that move in a scene is a fundamental computation performed by the visual system. This computation is greatly complicated by observer motion, which causes most objects to move across the retinal image.
How the visual system detects scene-relative object motion during self-motion is poorly understood.
Human behavioral studies suggest that the visual system may identify local conflicts between motion parallax and binocular disparity cues to depth and may use these signals to detect moving objects.
We describe a novel mechanism for performing this computation based on neurons in macaque middle temporal (MT) area with incongruent depth tuning for binocular disparity and motion parallax cues.
Neurons with incongruent tuning respond selectively to scene-relative object motion, and their responses are predictive of perceptual decisions when animals are trained to detect a moving object during self-motion.
This finding establishes a novel functional role for neurons with incongruent tuning for multiple depth cues.