New! Sign up for our free email newsletter.
Science News
from research organizations

Interactive cyber-physical human: Generating contact-rich whole-body motions

Date:
February 6, 2023
Source:
Tokyo University of Science
Summary:
Performing human-like motions that involve multiple contacts is challenging for robots. In this regard, a researcher has envisioned an interactive cyber-physical human (iCPH) platform with complementary humanoid (physical twin) and simulation (digital twin) elements. iCPH combines human measurement data, musculoskeletal analysis, and machine learning for data collection and augmentation. As a result, iCPH can understand, predict, and synthesize whole-body contact motions.
Share:
FULL STORY

Humans naturally perform numerous complex tasks. These include sitting down, picking something up from a table, and pushing a cart. These activities involve various movements and require multiple contacts, which makes it difficult to program robots to perform them.

Recently, Professor Eiichi Yoshida of the Tokyo University of Science has put forward the idea of an interactive cyber-physical human (iCPH) platform to tackle this problem. It can help understand and generate human-like systems with contact-rich whole-body motions. His work was published in Frontiers in Robotics and AI.

Prof. Yoshida briefly describes the fundamentals of the platform. "As the name suggests, iCPH combines physical and cyber elements to capture human motions. While a humanoid robot acts as a physical twin of a human, a digital twin exists as a simulated human or robot in cyberspace. The latter is modeled through techniques such as musculoskeletal and robotic analysis. The two twins complement each other."

This research raises several key questions. How can humanoids mimic human motion? How can robots learn and simulate human behaviors? And how can robots interact with humans smoothly and naturally? Prof. Yoshida addresses them in this framework. First, in the iCPH framework, human motion is measured by quantifying the shape, structure, angle, velocity, and force associated with the movement of various body parts. In addition, the sequence of contacts made by a human is also recorded. As a result, the framework allows the generic description of various motions through differential equations and the generation of a contact motion network upon which a humanoid can act.

Second, the digital twin learns this network via model-based and machine learning approaches. They are bridged together by the analytical gradient computation method. Continual learning teaches the robot simulation how to perform the contact sequence. Third, iCPH enriches the contact motion network via data augmentation and applies the vector quantization technique. It helps extract the symbols expressing the language of contact motion. Thus, the platform allows the generation contact motion in inexperienced situations. In other words, robots can explore unknown environments and interact with humans by using smooth motions involving many contacts.

In effect, the author puts forward three challenges. These pertain to the general descriptors, continual learning, and symbolization of contact motion. Navigating them is necessary for realizing iCPH. Once developed, the novel platform will have numerous applications.

"The data from iCPH will be made public and deployed to real-life problems for solving social and industrial issues. Humanoid robots can release humans from many tasks involving severe burdens and improve their safety, such as lifting heavy objects and working in hazardous environments. iCPH can also be used to monitor tasks performed by humans and help prevent work-related ailments. Finally, humanoids can be remotely controlled by humans through their digital twins, which will allow the humanoids to undertake large equipment installation and object transportation," says Prof. Yoshida, on the applications of iCPH.

Using the iCPH as ground zero and with the help of collaborations from different research communities, including robotics, artificial intelligence, neuroscience, and biomechanics, a future with humanoid robots is not far.


Story Source:

Materials provided by Tokyo University of Science. Note: Content may be edited for style and length.


Journal Reference:

  1. Eiichi Yoshida. Towards understanding and synthesis of contact-rich anthropomorphic motions through interactive cyber-physical human. Frontiers in Robotics and AI, 2022; 9 DOI: 10.3389/frobt.2022.1019523

Cite This Page:

Tokyo University of Science. "Interactive cyber-physical human: Generating contact-rich whole-body motions." ScienceDaily. ScienceDaily, 6 February 2023. <www.sciencedaily.com/releases/2023/02/230206104116.htm>.
Tokyo University of Science. (2023, February 6). Interactive cyber-physical human: Generating contact-rich whole-body motions. ScienceDaily. Retrieved December 21, 2024 from www.sciencedaily.com/releases/2023/02/230206104116.htm
Tokyo University of Science. "Interactive cyber-physical human: Generating contact-rich whole-body motions." ScienceDaily. www.sciencedaily.com/releases/2023/02/230206104116.htm (accessed December 21, 2024).

Explore More

from ScienceDaily

RELATED STORIES