Imitation Learning, Physical AI, TeleRobotics, Teleoperation

Single Arm + Glove

Source :  https://www.senseglove.com/project-rembrandt/

Dual Arms + Gloves

Source :  IEEE telepresence | TNO

Humanoid

Source :  https://www.i-botics.com/projects/xprize/

Virtual

Source :  Haption

Real world component

Source :  Haption

Imitation Learning

Source :  https://pressroom.toyota.com/

Industrial applications

Source :  Haption

Source :  Haption

High-quality sense-of-touch data is becoming a fundamental requirement for robotics imitation learning, especially as robots must accurately reproduce human skills. Modern robot learning from demonstration systems increasingly depend on detailed sense-of-touch information to generate reliable behavioral cloning in robotics and robust robot skill learning. As human-robot learning advances, high performance haptic feedback such as force feedback device enables more stable robot policy learning and improves the consistency of demonstration-based robotics in both real-world and simulated-world environments.

“Our research in robotics is aimed at amplifying people rather than replacing them,” said Gill Pratt, Chief Scientist for Toyota Motor Corporation. “This new teaching technique is both very efficient and produces very high performing behaviors, enabling robots to much more effectively amplify people in many ways. Previous state-of-the-art techniques to teach robots new behaviors were slow, inconsistent, inefficient, and often limited to narrowly defined tasks performed in highly constrained environments.”

Source :  https://pressroom.toyota.com/toyota-research-institute-unveils-breakthrough-in-teaching-robots-new-behaviors/

Advances in inverse reinforcement learning in robotics, generative models for robot imitation, and diffusion models for robot control are transforming how robots acquire new skills. Emerging foundation models for robotics, offline imitation learning, and policy cloning for robotics enable greater autonomy, while techniques such as kinesthetic teaching for robots, high-quality demonstration datasets for robotics, and robust robot motion imitation push real-world performance forward. These innovations are increasingly critical for industrial robotics imitation learning, surgical robotics, service robots, robotic assembly, human-robot collaboration across sectors.
They also contribute to the rise of physical AI: systems that tightly couple perception, reasoning, and embodied action in the real world. This includes multimodal world models for robots, simulation-to-real transfer, large-scale robot learning from internet and teleoperation data, and foundation models that integrate vision, language, and control. Such capabilities allow robots to generalize across tasks, interact safely with humans, adapt to unstructured environments.

Discover how Haption’s products make enable to work efficiently on these challenges.

Real-time data streaming up to 1 kHz

Data collection from a high-performance haptic force-feedback teleoperation system

Designed to be compatible with all industrial and humanoid robots

Single Arm and Dual Arms setup.

CONFIGURATION #1. DIY

For robotics, haptics, and real-time control experts.

Maximum flexibility and customization (C++, ROS, ROS2, …)

Full control over algorithms and system performance

Ideal for pushing the performance to the limits

CONFIGURATION #2. Plug & Play Standard or Customized

For AI, algorithm and real-time data collection experts.

Immediate access to real-time data : position, velocity, and force

Rapid deployment bundle with no complex integration

Focus entirely on algorithm development and imitation learning

Illustration of the “Plug & Play” configuration to enable real time data collection.
Others architectures are available, please contact us.

Adopt our technologies, accelerate your projects!