Physical AI Uses Both Sight and Touch to Manipulate Objects Like a Human
en-GBde-DEes-ESfr-FR

Physical AI Uses Both Sight and Touch to Manipulate Objects Like a Human

03/09/2025 Tohoku University

In everyday life, it's a no-brainer for us to grab a cup of coffee from the table. We seamlessly combine multiple sensory inputs such as sight (seeing how far away the cup is) and touch (feeling when our hand makes contact) in real-time without even thinking about it. However, recreating this in artificial intelligence (AI) is not quite as easy.

An international group of researchers created a new approach that integrates visual and tactile information to manipulate robotic arms, while adaptively responding to the environment. Compared to conventional vision-based methods, this approach achieved higher task success rates. These promising results represent a significant advancement in the field of multimodal physical AI.

Details of their breakthrough were published in the journal IEEE Robotics and Automation Letters on July 2, 2025.

Machine learning can be used to support artificial intelligence (AI) to learn human movement patterns, enabling robots to autonomously perform daily tasks such as cooking and cleaning. For example, ALOHA (A Low-cost Open-source Hardware System for Bimanual Teleoperation) is a system developed by Stanford University that enables the low-cost and versatile remote operation and learning of dual-arm robots. Both hardware and software are open source, so the research team was able to build upon this base.

However, these systems mainly rely on visual information only. Therefore, they lack the same tactile judgements a human could make, such as distinguishing the texture of materials or the front and back sides of objects. For example, it can be easier to tell which is the front or back side of Velcro by simply touching it instead of discerning how it looks. Relying solely on vision without other input is an unfortunate weakness.

"To overcome these limitations, we developed a system that also enables operational decisions based on the texture of target objects - which are difficult to judge from visual information alone," explains Mitsuhiro Hayashibe, a professor at Tohoku University's Graduate School of Engineering. "This achievement represents an important step toward realizing a multimodal physical AI that integrates and processes multiple senses such as vision, hearing, and touch - just like we do."

The new system was dubbed "TactileAloha." They found that the robot could perform appropriate bimanual operations even in tasks where front-back differences and adhesiveness are crucial, such as with Velcro and zip ties. They found that by applying vision-tactile transformer technology, their Physical AI robot exhibited more flexible and adaptive control.

The improved physical AI method was able to accurately manipulate objects, by combining multiple sensory inputs to form adaptive, responsive movements. There are nearly endless possible practical applications of these types of robots to lend a helping hand. Research contributions such as TactileAloha bring us one step closer to these robotic helpers becoming a seamless part of our everyday lives.

The research group was comprised of members from Tohoku University's Graduate School of Engineering and the Centre for Transformative Garment Production, Hong Kong Science Park, and the University of Hong Kong.

Title: TactileAloha: Learning Bimanual Manipulation with Tactile Sensing
Authors: Ningquan Gu, Kazuhiro Kosuge, Mitsuhiro Hayashibe
Journal: IEEE Robotics and Automation Letters
DOI: 10.1109/LRA.2025.3585396
Attached files
  • TactileAloha physical AI robot holding velcro. ©Tohoku University
  • Based on camera information, the arm grips both ends of the Velcro (A.1, B.1). Using tactile information, it senses the orientation of the tape and adjusts the posture and angle to align the hook surface with the loop surface (A.2, B.2). The Velcro is fixed, and the right arm presses it to ensure a firm connection (A.3, B.3). Different tape manipulation movements are automatically generated to adapt to the situation. ©Tohoku University
  • Video of the physical AI in action. ©Tohoku University
03/09/2025 Tohoku University
Regions: Asia, Japan, Hong Kong
Keywords: Applied science, Artificial Intelligence, Engineering, Technology

Disclaimer: AlphaGalileo is not responsible for the accuracy of content posted to AlphaGalileo by contributing institutions or for the use of any information through the AlphaGalileo system.

Testimonials

For well over a decade, in my capacity as a researcher, broadcaster, and producer, I have relied heavily on Alphagalileo.
All of my work trips have been planned around stories that I've found on this site.
The under embargo section allows us to plan ahead and the news releases enable us to find key experts.
Going through the tailored daily updates is the best way to start the day. It's such a critical service for me and many of my colleagues.
Koula Bouloukos, Senior manager, Editorial & Production Underknown
We have used AlphaGalileo since its foundation but frankly we need it more than ever now to ensure our research news is heard across Europe, Asia and North America. As one of the UK’s leading research universities we want to continue to work with other outstanding researchers in Europe. AlphaGalileo helps us to continue to bring our research story to them and the rest of the world.
Peter Dunn, Director of Press and Media Relations at the University of Warwick
AlphaGalileo has helped us more than double our reach at SciDev.Net. The service has enabled our journalists around the world to reach the mainstream media with articles about the impact of science on people in low- and middle-income countries, leading to big increases in the number of SciDev.Net articles that have been republished.
Ben Deighton, SciDevNet

We Work Closely With...


  • e
  • The Research Council of Norway
  • SciDevNet
  • Swiss National Science Foundation
  • iesResearch
Copyright 2025 by AlphaGalileo Terms Of Use Privacy Statement