“The most profound technologies are those that disappear”

 

Mark Weiser

Capturing Hand-Object Interaction and Reconstruction of Manipulated Objects

Image courtesy of Marc Mennigmann for the hands-project. Used after granted license. 

 

 

~~~ Download Link ~~~

 

Abstract

Hand motion capture with an RGB-D sensor gained recently a lot of research attention, however, even most recent approaches focus on the case of a single isolated hand. 

We focus instead on hands that interact with other hands or with a rigid or articulated object. 

 

Our framework successfully captures motion in such scenarios by combining a generative model with discriminatively trained salient points, collision detection and physics simulation to achieve a low tracking error with physically plausible poses. 

All components are unified in a single objective function that can be optimized with standard optimization techniques. 

We initially assume a-priori knowledge of the object's shape and skeleton. 

 

In case of unknown object shape there are existing 3d reconstruction methods that capitalize on distinctive geometric or texture features. 

These methods though fail for textureless and highly symmetric objects like household articles, mechanical parts or toys. 

We show that extracting 3d hand motion for in-hand scanning effectively facilitates the reconstruction of such objects and we fuse the rich additional information of hands into a 3d reconstruction pipeline. 

 

Finally, although shape reconstruction is enough for rigid objects, there is a lack of tools that build rigged models of articulated objects that deform realistically using RGB-D data. 

We propose a method that creates a fully rigged model consisting of a watertight mesh, embedded skeleton and skinning weights by employing a combination of deformable mesh tracking, motion segmentation based on spectral clustering and skeletonization based on mean curvature flow.