Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Seeing is believing: body motion dominates in multisensory conversations

Published: 26 July 2010 Publication History

Abstract

In many scenes with human characters, interacting groups are an important factor for maintaining a sense of realism. However, little is known about what makes these characters appear realistic. In this paper, we investigate human sensitivity to audio mismatches (i.e., when individuals' voices are not matched to their gestures) and visual desynchronization (i.e., when the body motions of the individuals in a group are mis-aligned in time) in virtual human conversers. Using motion capture data from a range of both polite conversations and arguments, we conduct a series of perceptual experiments and determine some factors that contribute to the plausibility of virtual conversing groups. We found that participants are more sensitive to visual desynchronization of body motions, than to mismatches between the characters' gestures and their voices. Furthermore, synthetic conversations can appear sufficiently realistic once there is an appropriate balance between talker and listener roles. This is regardless of body motion desynchronization or mismatched audio.

Supplementary Material

JPG File (tp084-10.jpg)
Supplemental material. (091.zip)
MP4 File (tp084-10.mp4)

References

[1]
Bernard, S., Therien, J., Malone, C., Beeson, S., Gubman, A., and Pardo, R., 2008. Taming the Mob: Creating believable crowds in Assassin's Creed. Presented at Game Developers Conference (San Francisco, CA, Feb 18--22).
[2]
Bickmore, T., and Cassell, J. 2005. Social Dialogue with Embodied Conversational Agents. In Advances in natural multimodal dialogue systems, 23--54.
[3]
Briton, N. J., and Hall, J. A. 1995. Beliefs about female and male nonverbal communication. Sex Roles: Journal of Research 32, 1--2, 79--90.
[4]
Cassell, J., Nakano, Y., Bickmore, T., Sidner, C., and Rich, C. 2001. Non-verbal cues for discourse structure. In Proceedings of the 39th Annual Meeting on Association for Computational Linguistics, 114--123.
[5]
Cassell, J., Vihjálmsson, H., and Bickmore, T. 2001. BEAT: the Behavior Expression Animation Toolkit. In Proceedings of SIGGRAPH 2001, 477--486.
[6]
Durupinar, F., Allbeck, J., Pelechano, N., and Badler, N. 2008. Creating crowd variation with the OCEAN personality model. In Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems, 1217--1220.
[7]
2010. Edge online: Performance art. http://www.edge-online.com/magazine/performance-art.
[8]
Ekman, P., and Friesen, W. 1969. The repertoire of nonverbal behavior: Categories, origins, usage, and coding. Semiotica 1, 1, 49--98.
[9]
Ekman, P. 1992. An argument for basic emotions. Cognition and Emotion 6, 3, 169--200.
[10]
Giorgolo, G., and Verstraten, F. 2008. Perception of speech-and-gesture integration. In Proceedings of the International Conference on Auditory-Visual Speech Processing 2008, 31--36.
[11]
Goldin-Meadow, S. 2005. Hearing gesture: How our hands help us think. Belknap Press.
[12]
Johansson, G. 1973. Visual perception of biological motion and a model for its analysis. Perception and Psychophysics 14, 2, 201--211.
[13]
Kendon, A. 1994. Do gestures communicate? A review. Research on language and social interaction 27, 3, 175--200.
[14]
Krahmer, E., and Swerts, M. 2007. The effects of visual beats on prosodic prominence: Acoustic analyses, auditory perception and visual perception. Journal of Memory and Language 57, 3, 396--414.
[15]
Lerner, A., Fitusi, E., Chrysanthou, Y., and Cohen-Or, D. 2009. Fitting behaviors to pedestrian simulations. In SCA '09: Proceedings of the 2009 ACM SIGGRAPH/Eurographics Symposium on Computer Animation, 199--208.
[16]
Levine, S., Theobalt, C., and Koltun, V. 2009. Real-time prosody-driven synthesis of body language. ACM Transactions on Graphics 28, 5, 1--10.
[17]
McDonnell, R., Larkin, M., Dobbyn, S., Collins, S., and O'Sullivan, C. 2008. Clone attack! perception of crowd variety. ACM Transactions on Graphics 27, 3, 26:1--26:8.
[18]
McDonnell, R., Ennis, C., Dobbyn, S., and O'Sullivan, C. 2009. Talking bodies: Sensitivity to desynchronization of conversations. ACM Transactions on Applied Perception 6, 4, 22:1--22:8.
[19]
McNeill, D. 1996. Hand and mind: What gestures reveal about thought. University of Chicago Press.
[20]
Musicant, A. D., and Butler, R. A. 1984. The influence of pinnae-based spectral cues on sound localization. The Journal of the Acoustical Society of America 75, 4, 1195--1200.
[21]
Neff, M., Kipp, M., Albrecht, I., and Seidel, H. 2008. Gesture modeling and animation based on a probabilistic re-creation of speaker style. ACM Transactions on Graphics 27, 1, 1--24.
[22]
Peters, C., and Ennis, C. 2009. Modeling groups of plausible virtual pedestrians. IEEE Computer Graphics and Applications 29, 4, 54--63.
[23]
Rose, D., and Clarke, T. J. 2009. Look who's talking: Visual detection of speech from whole-body biological motion cues during emotive interpersonal conversation. Perception 38, 1, 153--156.
[24]
Rouse, R. 1998. Embrace your limitations -- cut-scenes in computer games. ACM SIGGRAPH Computer Graphics 32, 4, 7--9.
[25]
Vilhjálmsson, H., and Cassell, J. 1998. Bodychat: autonomous communicative behaviors in avatars. In Proceedings of the second international conference on Autonomous agents, 269--276.

Cited By

View all
  • (2024)Resolving Collisions in Dense 3D Crowd AnimationsACM Transactions on Graphics10.1145/368726643:5(1-14)Online publication date: 6-Sep-2024
  • (2024)Audience perceptions of Foley footsteps and 3D realism designed to convey walker characteristicsPersonal and Ubiquitous Computing10.1007/s00779-024-01819-328:5(779-799)Online publication date: 1-Oct-2024
  • (2023)Data-Driven Communicative Behaviour Generation: A SurveyACM Transactions on Human-Robot Interaction10.1145/3609235Online publication date: 16-Aug-2023
  • Show More Cited By

Index Terms

  1. Seeing is believing: body motion dominates in multisensory conversations

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Graphics
    ACM Transactions on Graphics  Volume 29, Issue 4
    July 2010
    942 pages
    ISSN:0730-0301
    EISSN:1557-7368
    DOI:10.1145/1778765
    Issue’s Table of Contents
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 26 July 2010
    Published in TOG Volume 29, Issue 4

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. conversational agents
    2. crowds
    3. perception

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)16
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 18 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Resolving Collisions in Dense 3D Crowd AnimationsACM Transactions on Graphics10.1145/368726643:5(1-14)Online publication date: 6-Sep-2024
    • (2024)Audience perceptions of Foley footsteps and 3D realism designed to convey walker characteristicsPersonal and Ubiquitous Computing10.1007/s00779-024-01819-328:5(779-799)Online publication date: 1-Oct-2024
    • (2023)Data-Driven Communicative Behaviour Generation: A SurveyACM Transactions on Human-Robot Interaction10.1145/3609235Online publication date: 16-Aug-2023
    • (2023)Exploring the Appearance and Voice Mismatch of Virtual Characters2023 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct)10.1109/ISMAR-Adjunct60411.2023.00118(555-560)Online publication date: 16-Oct-2023
    • (2022)Multiparty Interaction Between Humans and Socially Interactive AgentsThe Handbook on Socially Interactive Agents10.1145/3563659.3563665(113-154)Online publication date: 27-Oct-2022
    • (2022)Manipulating Foley Footsteps and Character Realism to Influence Audience Perceptions of a 3D Animated Walk CycleProceedings of the 17th International Audio Mostly Conference10.1145/3561212.3561221(113-120)Online publication date: 6-Sep-2022
    • (2022)Motion and Meaning: Data-Driven Analyses of The Relationship Between Gesture and Communicative SemanticsProceedings of the 10th International Conference on Human-Agent Interaction10.1145/3527188.3561941(227-235)Online publication date: 5-Dec-2022
    • (2022)Labeling the Phrases of a Conversational Agent with a Unique Personalized Vocabulary2022 IEEE/SICE International Symposium on System Integration (SII)10.1109/SII52469.2022.9708605(856-863)Online publication date: 9-Jan-2022
    • (2022)The Handbook on Socially Interactive AgentsundefinedOnline publication date: 27-Oct-2022
    • (2021)A dataset of human and robot approach behaviors into small free-standing conversational groupsPLOS ONE10.1371/journal.pone.024736416:2(e0247364)Online publication date: 25-Feb-2021
    • Show More Cited By

    View Options

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media