Home > News > Touch-sensitive avatar-robotic arm based on real-time haptics November 7, 2017, Keio University

Touch-sensitive avatar-robotic arm based on real-time haptics November 7, 2017, Keio University

Touch-sensitive avatar-robotic arm based on real time haptics

Touch sensitive avatar-robotic arm based on real-time haptics developed by Takahiro Nozaki and colleagues at the Keio University Haptics Research Center. Credit: Keio University

Researchers at Keio University's Haptics Research Center have developed a 'real-time-avatar-robotic arm' that transmits sound, vision, and highly sensitive feelings of touch to remotely located users. This innovative touch sensitive robotic technology was reported in the October 2017 issue of IEEE Transactions on Industrial Electronics and demonstrated at CEATEC (October 2017, Tokyo). The technology is expected to find applications in areas such as industrial manufacturing, harvesting farm produce, and nursing care.

There is demand for to overcome the daunting challenges of the 21st century such as providing care for the elderly in rapidly aging industrialized nations, supporting labor intensive agriculture, and responding to extreme emergencies where humans cannot intervene directly, such as  disasters.
With this background, a growing number of researchers are focussing on the potential of 'haptics'— man-machine communication based on

 

—to solve these and related problems. In its simplest form, haptics enables users to feel the sense of touch via vibrations of forced motion. Such technology employs touch sensors that can be difficult to calibrate and often malfunction in extreme environments such as heat and radiation. Furthermore, conventional haptics technology is based on vibrations and is pseudo-tactile. So, although it can be used for games and entertainment, its range of industrial applications is very limited.

Takahiro Nozaki and colleagues of the Faculty of Science and Technology and Haptics Research Center at Keio University developed a haptic-based avatar-robot with a General Purpose Arm (GPA) that transmits sound, vision, movement, and importantly, highly sensitive sense of touch (force tactile transmission), to a remotely located user in real time. "This 'real haptics' is an integral part of the Internet of Actions (IoA) technology, having applications in manufacturing, agriculture, medicine, and ," says Nozaki.

Touch-sensitive avatar-robotic arm based on real time haptics

 

The robot pours water from a bottle into a cup. Credit: Keio University

This is the world's first high-precision tactile force transmission technology that remembers human movements, edits them, and reproduces them. Also, this arm does not employ conventional touch sensors, thereby making it cheaper, more compact, and robust with respect to malfunction and noise. The core technology behind this avatar-robot is based on high precision motors integrated in the avatar arm and algorithms to drive them. High precision control of force and position is critical for transmitting a sense of touch without using touch sensors.

Nozaki and colleagues have launched "Motion Lib' to commercialize their 'real-haptics ." The main product is an integrated chip called the "ABC-CORE' IC force/tactile controller. This IC chip controls the force adjustment of DC/AC servomotors and forces tactile transmission with two motors synchronized in motion. Importantly, since the load force applied to the motor is calculated by an algorithm in the chip, it is not necessary to install force or torque sensors.

High precision robotic arms are widely used in industry, for repetitive actions in automo technology bile assembly lines, for example. However, such robotic arms only repeat a preprogramed series of commands, grabbing well-defined, solid components used for constructing cars.

The challenge is to be able to recognize the shape, material composition—soft or hard— and position of an object, and manipulate it according to real-time instructions from a user located at a distance from the arm, where the arm acts as a real-time avatar.

https://youtu.be/oXNyUMPiHFE

The critical technical breakthroughs in motor control and robotics for the robotic-avatar developed by Nozaki and co-workers were first reported by Keio University's Kouhei Ohnishi, in 1983 in a paper titled, "Torque –speed regulation of DC motor based on load torque estimation method."

Ohnishi continued to develop his ideas in his 1993 paper on 'sensorless torque control' (IEEE Transactions on Industrial Electronics, 40, 259, (1993)).

This report was followed by his proposals for 'motion control in mechatronics' (IEEE Transactions on Mechatronics, 1, 56, (1996)).

Then, in 2004, Ohnishi addressed the issue of 'good sense of remote objects' in AMC 2004– Kawasaki, Japan.

Source:PHY

For more information, please visit our website:www.qualitymaterial.net,

send us email at angel.ye @powerwaywafer.com or powerwaymaterial@gmail.com.