Simulation-Ready Hair Capture

 

We present the first method for capturing dynamic hair and automatically determining the physical properties for simulating the observed hairstyle in motion.

April 24, 2017
Eurographics 2017

 

Authors

Liwen Hu (Disney Research/Pinscreen/University of Southern California)

Derek Bradley (Disney Research)

Hao Li (Pinscreen/University of Southern California)

Thabo Beeler (Disney Research)

Simulation-Ready Hair Capture

Abstract

Physical simulation has long been the approach of choice for generating realistic hair animations in CG. A constant drawback of simulation, however, is the necessity to manually set the physical parameters of the simulation model in order to get the desired dynamic behavior. To alleviate this, researchers have begun to explore methods for reconstructing hair from the real world and even to estimate the corresponding simulation parameters through the process of inversion. So far, however, these methods have had limited applicability, because dynamic hair capture can only be played back without the ability to edit, and solving for simulation parameters can only be accomplished for static hairstyles, ignoring the dynamic behavior. We present the first method for capturing dynamic hair and automatically determining the physical properties for simulating the observed hairstyle in motion. Since our dynamic inversion is agnostic to the simulation model, the proposed method applies to virtually any hair simulation technique, which we demonstrate using two state-of-the-art hair simulation models. The output of our method is a fully simulation-ready hairstyle, consisting of both the static hair geometry as well as its physical properties. The hairstyle can be easily edited by adding additional external forces, changing the head motion, or re-simulating in completely different environments, all while remaining faithful to the captured hairstyle.

Copyright Notice