From touch to vision: A bioinspired multisensory framework brings human-like perception to robots
en-GBde-DEes-ESfr-FR

From touch to vision: A bioinspired multisensory framework brings human-like perception to robots

27/04/2026 TranSpread

In humans, sensory information is not processed in isolation. Vision, touch, hearing, smell, and taste interact dynamically to form unified perceptions of the environment. This cross-modal integration enables rapid decision-making, contextual understanding, and even imagination. However, most artificial intelligence systems still rely on isolated sensory channels or energy-intensive centralized processing. Existing multisensory approaches often struggle with scalability, adaptability, and power efficiency, limiting their application in autonomous robots and embodied intelligence. Moreover, artificial systems rarely achieve true cross-modal reconfiguration, where information from one sense can be transformed into another. Based on these challenges, there is a pressing need to develop an energy-efficient multisensory framework capable of human-like cross-modal cognition.

Researchers from the Beijing Institute of Nanoenergy and Nanosystems, the University of Chinese Academy of Sciences, Guangxi University, and Georgia Institute of Technology report a bioinspired multisensory system that closely mimics how the human brain integrates information across senses. Published (DOI: 10.1016/j.esci.2025.100482) online on March 2026, in eScience, the study introduces a triboelectric-driven framework that unifies visual, tactile, auditory, olfactory, and gustatory inputs within a self-powered architecture. The system demonstrates high-accuracy cross-modal recognition and adaptive sensory reconfiguration, offering a new pathway toward energy-autonomous robotic cognition.

At the core of the framework is a triboelectric-sensing-mediated artificial neural network (TES-ANN), inspired by the distributed and hierarchical organization of human sensory neurons. Mechanical, acoustic, and material-based stimuli are converted into electrical signals through triboelectric nanogenerators, eliminating the need for external power supplies. These signals are then encoded as neural-like spikes and processed through interconnected artificial sensory modules.

The system demonstrates robust tactile-visual association: handwritten digits and letters sensed through touch are reconstructed as visual images with an accuracy of 97.12%. Similarly, auditory inputs are successfully linked to corresponding visual, olfactory, and gustatory representations, achieving 94.62% accuracy in cross-modal reconfiguration. Beyond empirical learning, the framework exhibits non-empirical cognitive behavior. For example, after learning basic associations between colors and fruits, the system can infer and generate a plausible visual representation of a previously unseen object—such as a “purple strawberry”—based solely on auditory input.

This ability to move from perception to inference and imagination marks a conceptual shift from passive multisensory fusion to active cross-modal cognition. Importantly, all these functions are achieved with high energy efficiency, addressing a key bottleneck in autonomous intelligent systems.

“This work demonstrates that artificial systems can move beyond simple sensory recognition toward genuine cognitive association and imagination,” said the study’s corresponding author. “By combining triboelectric sensing with bioinspired neural architectures, we show that energy-autonomous systems can perform complex cross-modal tasks once thought to be exclusive to biological brains. This opens new opportunities for developing intelligent machines that interact with the world in a far more natural and adaptive way.”

The proposed multisensory framework has broad implications for robotics, human–machine interfaces, and embodied artificial intelligence. Energy-autonomous robots equipped with such systems could operate for extended periods in complex environments, interpreting sensory cues more holistically and responding more intelligently. Potential applications range from assistive prosthetics and rehabilitation devices to immersive virtual-reality systems and camera-free object recognition. By enabling machines to associate, infer, and even imagine across sensory domains, this research lays the groundwork for a new generation of intelligent systems that more closely resemble human cognition, bridging the gap between perception and understanding.

###

References

DOI

10.1016/j.esci.2025.100482

Original Source URL

https://doi.org/10.1016/j.esci.2025.100482

Funding information

This work is supported by the National Key Research and Development Program of China (2023YFB3208102), the National Natural Science Foundation of China (52073031), the “Hundred Talents Program” of the Chinese Academy of Science.

About eScience

eScience – a Diamond Open Access journal cooperated with KeAi and published online at ScienceDirect. eScience is founded by Nankai University (China) in 2021 and aims to publish high quality academic papers on the latest and finest scientific and technological research in interdisciplinary fields related to energy, electrochemistry, electronics, and environment. eScience provides insights, innovation and imagination for these fields by built consecutive discovery and invention. Now eScience has been indexed by SCIE, CAS, Scopus and DOAJ. Its impact factor is 36.6, which is ranked first in the field of electrochemistry.

Paper title: Bioinspired triboelectric-driven multisensory framework with autonomous cross-modal adaptation
Attached files
  • Bioinspired triboelectric-driven multisensory framework for cross-modal associative learning. Schematic illustration of a bioinspired multisensory framework enabled by triboelectric sensing and artificial neural networks (TES-ANN). Signals from vision, touch, hearing, smell, and taste are first captured by distributed triboelectric sensors and encoded into neural-like electrical spikes. These multimodal signals are transmitted and modulated before being integrated within an artificial neural network that performs cross-modal associative learning. Through this process, information from one sensory modality can be reconfigured into another, enabling recognition, inference, and imagination in a self-powered and energy-efficient manner, mimicking key features of human multisensory cognition.
27/04/2026 TranSpread
Regions: North America, United States, Asia, China
Keywords: Applied science, Technology

Disclaimer: AlphaGalileo is not responsible for the accuracy of content posted to AlphaGalileo by contributing institutions or for the use of any information through the AlphaGalileo system.

Testimonials

For well over a decade, in my capacity as a researcher, broadcaster, and producer, I have relied heavily on Alphagalileo.
All of my work trips have been planned around stories that I've found on this site.
The under embargo section allows us to plan ahead and the news releases enable us to find key experts.
Going through the tailored daily updates is the best way to start the day. It's such a critical service for me and many of my colleagues.
Koula Bouloukos, Senior manager, Editorial & Production Underknown
We have used AlphaGalileo since its foundation but frankly we need it more than ever now to ensure our research news is heard across Europe, Asia and North America. As one of the UK’s leading research universities we want to continue to work with other outstanding researchers in Europe. AlphaGalileo helps us to continue to bring our research story to them and the rest of the world.
Peter Dunn, Director of Press and Media Relations at the University of Warwick
AlphaGalileo has helped us more than double our reach at SciDev.Net. The service has enabled our journalists around the world to reach the mainstream media with articles about the impact of science on people in low- and middle-income countries, leading to big increases in the number of SciDev.Net articles that have been republished.
Ben Deighton, SciDevNet

We Work Closely With...


  • The Research Council of Norway
  • SciDevNet
  • Swiss National Science Foundation
  • iesResearch
Copyright 2026 by AlphaGalileo Terms Of Use Privacy Statement