Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Leveraging motion capture and 3D scanning for high-fidelity facial performance acquisition

Published: 25 July 2011 Publication History

Abstract

This paper introduces a new approach for acquiring high-fidelity 3D facial performances with realistic dynamic wrinkles and fine-scale facial details. Our approach leverages state-of-the-art motion capture technology and advanced 3D scanning technology for facial performance acquisition. We start the process by recording 3D facial performances of an actor using a marker-based motion capture system and perform facial analysis on the captured data, thereby determining a minimal set of face scans required for accurate facial reconstruction. We introduce a two-step registration process to efficiently build dense consistent surface correspondences across all the face scans. We reconstruct high-fidelity 3D facial performances by combining motion capture data with the minimal set of face scans in the blendshape interpolation framework. We have evaluated the performance of our system on both real and synthetic data. Our results show that the system can capture facial performances that match both the spatial resolution of static face scans and the acquisition speed of motion capture systems.

Supplementary Material

Supplemental material. (a74-huang.zip)
MP4 File (tp074_11.mp4)

References

[1]
Alexa, M. 2003. Differential coordinates for local mesh morphing and deformation. The Visual Computer. 19(2):105--114.
[2]
Alexander, O., Rogers, M., Lambeth, W., Chiang, M., and Debevec, P. 2009. The digital emily project: photo-real facial modeling and animation. In ACM SIGGRAPH 2009 Courses, 12:1--12:15.
[3]
Beeler, T., Bickel, B., Beardsley, P., Sumner, B., and Gross, M. 2010. High-quality single-shot capture of facial geometry. ACM Trans. Graph. 29, 4, 40:1--40:9.
[4]
Bickel, B., Botsch, M., Angst, R., Matusik, W., Otaduy, M., Pfister, H., and Gross, M. 2007. Multi-scale capture of facial geometry and motion. ACM Trans. Graph. 26, 3, 33:1--33:10.
[5]
Blanz, V., Basso, C., Poggio, T., and Vetter, T. 2003. Reanimating faces in images and video. In Computer Graphics Forum. 22(3):641--650.
[6]
Borshukov, G., Piponi, D., Larsen, O., Lewis, J. P., and Tempelaar-Lietz, C. 2003. Universal capture: image-based facial animation for "the matrix reloaded". In ACM SIGGRAPH 2003 Sketches & Applications, 1--1.
[7]
Bradley, D., Heidrich, W., Popa, T., and Sheffer, A. 2010. High resolution passive facial performance capture. ACM Trans. Graph. 29, 4, 41:1--41:10.
[8]
Cazals, F., and Pouget, M. 2003. Estimating differential quantities using polynomial fitting of osculating jets. In Proceedings of the 2003 Eurographics/ACM SIGGRAPH symposium on Geometry processing, 177--187.
[9]
Chai, J., and Hodgins, J. 2007. Constraint-based motion optimization using a statistical dynamic model. ACM Transactions on Graphics 26, 3, 8:1--8:9.
[10]
Chai, J., Xiao, J., and Hodgins, J. 2003. Vision-based control of 3D facial animation. In Proceedings of the 2003 ACM SIGGRAPH/Eurographics Symposium on Computer Animation. 193--206.
[11]
DeCarlo, D., and Metaxas, D. 2000. Optical flow constraints on deformable models with applications to face tracking. International Journal of Computer Vision. 38(2):99--127.
[12]
Essa, I., Basu, S., Darrell, T., and Pentland, A. 1996. Modeling, tracking and interactive animation of faces and heads using input from video. In Proceedings of Computer Animation Conference. 68--79.
[13]
Guenter, B., Grimm, C., Wood, D., Malvar, H., and Pighin, F. 1998. Making Faces. In Proceedings of ACM SIGGRAPH 1998, 55--66.
[14]
James, D. L., and Twigg, C. D. 2005. Skinning mesh animations. ACM Transactions on Graphics 24, 3, 399--407.
[15]
Lau, M., Chai, J., Xu, Y.-Q., and Shum, H. 2009. Face poser: Interactive modeling of 3d facial expressions using facial priors. ACM Transactions on Graphics 29, 1, 3:1--3:17.
[16]
Li, H., Adams, B., Guibas, L. J., and Pauly, M. 2009. Robust single-view geometry and motion reconstruction. ACM Trans. Graph. 28, 5, 175:1--175:10.
[17]
Li, H., Weise, T., and Pauly, M. 2010. Example-based facial rigging. ACM Transactions on Graphics 29, 4, 32:1--32:6.
[18]
Ma, W.-C., Jones, A., Chiang, J.-Y., Hawkins, T., Frederiksen, S., Peers, P., Vukovic, M., Ouhyoung, M., and Debevec, P. 2008. Facial performance synthesis using deformation-driven polynomial displacement maps. ACM Trans. Graph. 27, 5, 121:1--121:10.
[19]
Papenberg, N., Bruhn, A., Brox, T., Didas, S., and Weickert, J. 2006. Highly accurate optic flow computation with theoretically justified warping. Int. J. Comput. Vision 67, 2, 141--158.
[20]
Pighin, F., and Lewis, J. P. 2006. Facial motion retargeting. In ACM SIGGRAPH 2006 Courses, SIGGRAPH '06.
[21]
Pighin, F., Szeliski, R., and Salesin, D. 1999. Resynthesizing facial animation through 3D model-based tracking. In International Conference on Computer Vision. 143--150.
[22]
Sorkine, O., Cohen-Or, D., Lipman, Y., Alexa, M., Rössl, C., and Seidel, H.-P. 2004. Laplacian surface editing. In SGP '04: Proceedings of the 2004 Eurographics/ACM SIGGRAPH symposium on Geometry processing, 175--184.
[23]
Sumner, R. W., Schmid, J., and Pauly, M. Embedded deformation for shape manipulation. ACM Trans. Graph. 26, 3, 80:1--80:7.
[24]
Vicon Systems, 2011. http://www.vicon.com.
[25]
Vlasic, D., Brand, M., Pfister, H., and Popović, J. 2005. Face transfer with multilinear models. ACM Transactions on Graphics 24, 3, 426--433.
[26]
Williams, L. 1990. Performance driven facial animation. In Proceedings of ACM SIGGRAPH 1990. 24(4):235--242.
[27]
XYZ RGB Systems, 2011. http://www.xyzrgb.com/.
[28]
Zhang, L., Snavely, N., Curless, B., and Seitz, S. 2004. Spacetime faces: high resolution capture for modeling and animation. ACM Transactions on Graphics 23, 3, 548--558.

Cited By

View all

Recommendations

Reviews

Fernando Santos Osorio

Huang et al. present "a new approach for acquiring high-fidelity [and realistic] 3D facial performances" (motion capture of facial expressions and facial movements). According to the authors, their approach "leverages state-of-the-art motion capture technology and advanced 3D scanning technology for facial performance acquisition," reducing the time and effort required for 3D facial acquisition. They propose a hybrid system combining high spatial resolution face scans (using a laser scanner) and high temporal resolution motion capture data (using a marker-based motion capture system). In their experiments, they use a 12-camera Vicon motion capture system to record high-resolution dynamic facial movements, placing about 100 retro-reflective markers on the face and setting the acquisition rate to 240 frames per second. They also use a Minolta VIVID 910 laser scanner to record the high-resolution static facial geometry of an actor, which acquires a face mesh with 100k to 200k vertices in about 2.5 seconds and achieves an accuracy of 0.008 mm. The combination of both marker-based face motion capture and high-resolution static facial data enables the capture of high-quality facial expressions. The method is based on the following steps. First, the authors record facial performances of an actor using optical motion capture systems. Second, they perform facial analysis on the captured data and obtain a minimal set of face scans required for accurate facial reconstruction. Third, they register motion-capture markers to all of the face scans and build consistent dense surface correspondences across all of the face scans. Finally, they combine motion-capture data with the minimal set of the face scans to reconstruct high-fidelity facial performances in the blendshape interpolation framework. The facial data analysis (step two) automatically selects a minimum set of facial expressions by minimizing the reconstruction errors associated with the selected facial expressions. Then, the marker mesh registration (step three) is applied to register motion capture data between the two capturing systems. The authors also propose the use of a region-based fine-scale mesh registration process. The final results, presented in the tests conducted by the authors, acquire 3D facial performances of three subjects and demonstrate that the system can capture realistic dynamic wrinkles and fine-scale facial details with good performance. The results are best seen in the accompanying video of this paper (as the authors indicate). Online Computing Reviews Service

Access critical reviews of Computing literature here

Become a reviewer for Computing Reviews.

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Graphics
ACM Transactions on Graphics  Volume 30, Issue 4
July 2011
829 pages
ISSN:0730-0301
EISSN:1557-7368
DOI:10.1145/2010324
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 25 July 2011
Published in TOG Volume 30, Issue 4

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. blendshape interpolation
  2. face modeling
  3. facial animation
  4. facial data analysis
  5. motion capture
  6. nonrigid surface registration

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)5
  • Downloads (Last 6 weeks)0
Reflects downloads up to 25 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Medial hex-meshing: high-quality all-hexahedral mesh generation based on medial meshEngineering with Computers10.1007/s00366-023-01925-540:4(2537-2557)Online publication date: 1-Aug-2024
  • (2024)On why mesh untangling may not be requiredEngineering with Computers10.1007/s00366-023-01913-940:3(1357-1374)Online publication date: 1-Jun-2024
  • (2024)An enhanced approach for inner and outer faces recognition of complex thin-shell partsEngineering with Computers10.1007/s00366-023-01802-140:1(493-525)Online publication date: 1-Feb-2024
  • (2022)Hex-Mesh Generation and Processing: A SurveyACM Transactions on Graphics10.1145/355492042:2(1-44)Online publication date: 18-Oct-2022
  • (2022)S3-SlicerACM Transactions on Graphics10.1145/3550454.355551641:6(1-15)Online publication date: 30-Nov-2022
  • (2022)DSG-Net: Learning Disentangled Structure and Geometry for 3D Shape GenerationACM Transactions on Graphics10.1145/352621242:1(1-17)Online publication date: 12-Aug-2022
  • (2022)A Large-Scale Comparison of Tetrahedral and Hexahedral Elements for Solving Elliptic PDEs with the Finite Element MethodACM Transactions on Graphics10.1145/350837241:3(1-14)Online publication date: 7-Mar-2022
  • (2022)Plausible 3D Face Wrinkle Generation Using Variational AutoencodersIEEE Transactions on Visualization and Computer Graphics10.1109/TVCG.2021.305125128:9(3113-3125)Online publication date: 1-Sep-2022
  • (2022)Transferring Facial Expressions to Digital Character2022 29th International Conference on Systems, Signals and Image Processing (IWSSIP)10.1109/IWSSIP55020.2022.9854433(1-4)Online publication date: 1-Jun-2022
  • (2022)SimuMan: A Simultaneous Real-Time Method for Representing Motions and Emotions of Virtual Human in MetaverseInternet of Things – ICIOT 202110.1007/978-3-030-96068-1_6(77-89)Online publication date: 18-Feb-2022
  • Show More Cited By

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media