Multimodal Contact Modeling - Honda Research Institute USA

Multimodal Contact Modeling

Multimodal Contact Modeling

Multimodal Contact Modeling

A multimodal perception system infers an object’s full 6D pose of object in contact, enabling reliable, human-like in-hand manipulation.

Determining an object’s precise position and orientation while it’s already inside a robotic hand is one of in-hand manipulation’s most difficult challenges. Cameras can’t “see” through fingers. IRR is addressing this with a multimodal perception system that fuses vision, depth, and tactile sensing.

Using a deep neural network trained on combined sensory data, the hand can infer the object’s full 6D pose throughout its manipulation, even for unseen grasps and unseen object orientations. This enables reliable in-hand manipulation, fine object reorientation, and transitions between grasp types without requiring a clear visual line of sight.

By closing this perception gap, we unlock more human-like dexterity and more capable remote operation workflows.

Learn more in the associated research publication.