The Brain’s Ability to Grasp the “Gist” of a Visual Scene Begins Earlier Than Expected
en-GBde-DEes-ESfr-FR

The Brain’s Ability to Grasp the “Gist” of a Visual Scene Begins Earlier Than Expected


When animals move through complex visual environments, the brain cannot afford to analyze every detail one by one. Instead, it rapidly extracts the overall structure of the scene—for example, the mean (average) direction of motion across many moving elements. This ability, known as ensemble perception, allows the brain to capture the “gist” of a scene at a glance. Yet where, and how, this statistical summary is computed in the brain has remained unclear.

A research team led by LEE Doyun and KIM Yee-Joon at the Center for Memory and Glioscience within the Institute for Basic Science (IBS) has now shown that this process begins much earlier in the visual system than previously thought.

Co-corresponding author LEE Doyun said, “What is especially striking is that this transformation begins already in primary visual cortex. The brain starts compressing complex sensory input into useful statistical summaries at a very early stage.”

In the brain, visual information is processed step by step along a hierarchy of regions. The primary visual cortex (V1) is the first cortical stage that receives visual input from the eyes and is traditionally thought to process simple features such as edges or motion direction. Further downstream, the posterior parietal cortex (PPC) integrates this information into more abstract representations that are linked to perception and decision-making.

The researchers found that V1 already encodes not only the mean direction of complex motion patterns, but also their variance—how dispersed or uncertain the motion is. This information is then carried forward to PPC, where it is reorganized into more abstract category representations that can guide behavior.

To investigate how the brain extracts these visual summaries, the team trained head-fixed mice to classify random-dot motion stimuli according to their overall direction. Unlike conventional motion displays, in which many dots move coherently in a single direction, the stimuli in this study were designed so that each dot moved in a different direction sampled from a controlled distribution. This allowed the researchers to independently manipulate the mean motion direction and its variability.

The mice successfully learned to group eight possible mean motion directions into two motion categories. Even when the motion of individual dots varied widely, the animals could still categorize the overall direction, indicating that they were not simply following a few prominent local signals. Instead, they were extracting a true statistical summary of the scene.

“We showed that the brain does not process complex visual input by tracking each element individually,” said LEE Young-Beom, first author of the study. “Instead, it extracts stable summary information such as mean and variance to rapidly capture the overall structure of the environment.”

Using miniscope calcium imaging, the researchers recorded neural activity in both V1 and PPC while the mice performed the task or passively viewed the stimuli. At the level of individual neurons, only a relatively small subset showed clear selectivity for the global mean motion direction. At the population level, however, neural activity in both regions robustly encoded the mean motion direction—even though most single neurons did not appear strongly tuned on their own.

The study revealed a clear division of labor across the cortical hierarchy. In V1, population activity encoded both the mean and the variance of motion direction, indicating that early visual cortex already computes summary statistics rather than merely relaying local signals. In PPC, by contrast, the representation shifted toward more abstract, task-relevant category information, suggesting that sensory summaries are progressively transformed into task-relevant signals.

The researchers also found that task demands could reshape early visual representations. During active categorization, the neural representation of mean motion direction in V1 became systematically biased toward the center of the learned category. This suggests that even early visual cortex is not purely stimulus-driven, but can be influenced by learning and behavioral context.

Another notable finding was that seemingly “untuned” neurons still contributed substantially to the population code. Even neurons that did not meet conventional selectivity criteria helped support accurate representation of global motion direction when analyzed collectively, highlighting the importance of distributed population coding in the brain.

Co-corresponding author KIM Yee-Joon added, “Our findings suggest that visual information is progressively reorganized—from summary statistics in early visual cortex to more abstract category representations in higher cortical areas. This provides an important clue to how the brain efficiently makes sense of complex scenes.”

By revealing how the brain converts noisy sensory input into stable statistical summaries and then into abstract category signals, the study provides new insight into a fundamental principle of perception. The findings may help explain how the brain rapidly extracts meaningful structure from complex environments and could also inform future work in artificial intelligence and computer vision.

The study was published online in Advanced Science on March 23, 2026.

- References

Young-Beom Lee, Oliver James, Gaeun Jung, Doyun Lee and Yee-Joon Kim. Hierarchical Summary Statistics Encoding Across Primary Visual and Posterior Parietal Cortices. Advanced Science.

DOI: 10.1002/advs.202512369
Archivos adjuntos
  • Figure 1. This study shows that mouse V1 simultaneously encodes the ensemble mean and variance of motion, providing a robust summary‐statistic representation that persists despite single-neuron variability. These signals propagate to PPC, where they are transformed into abstract category representations during decision making. The findings reveal a hierarchical compression of sensory detail into increasingly abstract statistical codes.
  • Figure 2. Experimental design, stimuli, and behavioral performancea. Schematic of the two alternative forced choice (2-AFC) motion categorization task. The head-fixed mice were trained to group eight global motion directions into the two discrete categories by turning the wheel to the left or right. We used eight global motion directions from the predefined set of directions from 22.5° to 337.5° in steps of 45°.b. Illustrations of how stimuli and conditions were created. First, one motion direction was sampled from eight possible motion directions and then used as a mean motion direction in each trial of two conditions. In homogeneous random dot kinematograms (RDKs), the sampled motion direction was identical to the motion direction of the rest of dots. In heterogeneous RDKs, each dot’s motion direction was randomly selected from a uniform distribution within one of the three ranges (90°, 180° and 270°) with one of eight global mean motion directions. Note that not a single dot in heterogeneous RDK was assigned to the global motion direction along the distribution’s mean.c. The slope values of cumulative normal functions fit to each mouse’s behavioral data are shown for the homogeneous and heterogeneous conditions. Error bars indicate ±SEM. Mice performed the task reliably up to a certain difficulty level in both homogeneous and heterogeneous stimulus conditions. A difference in task performance between the two conditions was observed only for the 270
  • Figure 3. Population-level representational analysis of V1 and PPC using miniscope calcium imaginga. Recording location of V1 and a representative calcium imaging image.b. RSA results in V1, showing dominant representation of mean direction information.c. Recording location of PPC and a representative calcium imaging image.d. RSA results in PPC, showing stronger representation of category information relative to mean direction information.Together, these findings suggest a hierarchical transformation of sensory summary information into increasingly abstract, behaviorally relevant category representations.
Regions: Asia, South Korea
Keywords: Science, Life Sciences

Disclaimer: AlphaGalileo is not responsible for the accuracy of content posted to AlphaGalileo by contributing institutions or for the use of any information through the AlphaGalileo system.

Testimonios

We have used AlphaGalileo since its foundation but frankly we need it more than ever now to ensure our research news is heard across Europe, Asia and North America. As one of the UK’s leading research universities we want to continue to work with other outstanding researchers in Europe. AlphaGalileo helps us to continue to bring our research story to them and the rest of the world.
Peter Dunn, Director of Press and Media Relations at the University of Warwick
AlphaGalileo has helped us more than double our reach at SciDev.Net. The service has enabled our journalists around the world to reach the mainstream media with articles about the impact of science on people in low- and middle-income countries, leading to big increases in the number of SciDev.Net articles that have been republished.
Ben Deighton, SciDevNet
AlphaGalileo is a great source of global research news. I use it regularly.
Robert Lee Hotz, LA Times

Trabajamos en estrecha colaboración con...


  • e
  • The Research Council of Norway
  • SciDevNet
  • Swiss National Science Foundation
  • iesResearch
Copyright 2026 by DNN Corp Terms Of Use Privacy Statement