Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3577190.3614169acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
research-article
Open access

On Head Motion for Recognizing Aggression and Negative Affect during Speaking and Listening

Published: 09 October 2023 Publication History

Abstract

Affective aggression is a form of aggression characterized by impulsive reactions driven by strong negative emotions. Despite the extensive research in the area of automatic emotion recognition, affective aggression is a phenomenon that has received less attention. This study investigates the use of head motion as a potential indicator of affective aggression and negative affect. It provides an analysis of head movement patterns associated with various levels of aggression, valence, arousal and dominance, and compares behaviors and recognition performance under speaking and listening conditions. The study was conducted on the Negative Affect and Aggression database - a multimodal corpus of dyadic interactions between aggression regulation training actors and non-actors, annotated for levels of aggression, valence, arousal, and dominance. Results demonstrate that head motion features can serve as promising indicators of affect during both speaking and listening. Valence and arousal prediction achieved better performance during speaking, while aggression and dominance were better predicted during listening. Significant increases in the magnitude of pitch angular acceleration were associated with escalation along all four annotated dimensions. Interestingly, higher escalation was accompanied by a significant increase in the total number of movements during speaking, but a significant decrease of the number of movements was observed as escalation increased along listening intervals. These findings are particularly relevant as head motion can be used solely or potentially as a supplementary modality when other modalities such as speech or facial expressions are unavailable or altered.

References

[1]
A. Adams, M. Mahmoud, T. Baltrusaitis, and P. Robinson. 2015. Decoupling facial expressions and head motions in complex emotions. In Proc. of on Affective Computing and Intelligent Interaction 2015. IEEE Computer Society, Washington, DC, USA, 274–280.
[2]
S. Alghowinem, R.d Goecke, M. Wagner, G. Parker, and M. Breakspear. 2013. Head pose and movement analysis as an indicator of depression. In Humaine Association Conf. on Affective Computing and Intelligent Interaction (ACII). IEEE Computer Society, Washington, DC, USA, 283–288.
[3]
Johnie J Allen and Craig A Anderson. 2017. Aggression and violence: Definitions and distinctions. In The Wiley handbook of violence and aggression. John Wiley & Sons, Ltd, Chichester, UK, 1–14. https://doi.org/10.1002/9781119057574.whbva001
[4]
Z. Ambadar, J.F. Cohn, and L.I. Reed. 2009. All Smiles are Not Created Equal: Morphology and Timing of Smiles Perceived as Amused, Polite, and Embarrassed/Nervous. J. of Nonverbal Behavior 33, 1 (2009), 17–34.
[5]
C. A. Anderson. 2000. Violence and aggression. In Encyclopedia of psychology, Vol. 8.American Psychological Association, Washington, DC, US, 162–169.
[6]
Craig A. Anderson and Brad J. Bushman. 2002. Human Aggression. Annual Review of Psychology 53, 1 (2002), 27–51.
[7]
J.-Y. Antoine, J. Villaneau, and A. Lefeuvre. 2014. Weighted Krippendorff’s alpha is a more reliable metrics for multi-coders ordinal annotations: experimental studies on emotion, opinion and coreference annotation. In EACL 2014. Association for Computational Linguistics, NY, USA, 10–p.
[8]
Gustavo E. A. P. A. Batista, Ronaldo C. Prati, and Maria Carolina Monard. 2004. A Study of the Behavior of Several Methods for Balancing Machine Learning Training Data. SIGKDD Explor. Newsl. 6, 1 (2004), 20–29. https://doi.org/10.1145/1007730.1007735
[9]
Maciej Behnke, Nadia Bianchi-Berthouze, and Lukasz D Kaczmarek. 2021. Head movement differs for positive and negative emotions in video recordings of sitting individuals. Scientific Reports 11, 1 (2021), 7405.
[10]
Kristin P. Bennett and Colin Campbell. 2000. Support Vector Machines: Hype or Hallelujah?SIGKDD Explor. Newsl. 2, 2 (2000), 1–13.
[11]
Leonard Berkowitz. 1998. 3 - Affective Aggression: The Role of Stress, Pain, and Negative Affect. In Human Aggression, Russell G. Geen and Edward Donnerstein (Eds.). Academic Press, San Diego, 49–72.
[12]
Ekaba Bisong. 2019. Logistic Regression. Apress, Berkeley, CA, 243–250.
[13]
S.M. Boker, J.F. Cohn, B.-J. Theobald, I. Matthews, T. Brick, and J.R. Spies. 2009. Effects of damping head movement and facial expression in dyadic conversation using real–time facial expression tracking and synthesized avatars. Philosophical Trans. of the Royal Society B: Biological Sciences 364, 1535 (2009), 3485–3495.
[14]
L. Breiman. 2001. Random Forests. Machine Learning 45, 1 (2001), 5–32.
[15]
Carlos Busso and Shrikanth S Narayanan. 2008. Scripted dialogs versus improvisation: Lessons learned about emotional elicitation techniques from the IEMOCAP database. In Proc. of Ninth annual conference of the international speech communication association, Interspeech 2008. ISCA, 1670–1673. https://doi.org/10.21437/Interspeech.2008-463
[16]
A. Carroll, M. McCarthy, S. Houghton, E. Sanders O’Connor, and C. Zadow. 2018. Reactive and proactive aggression as meaningful distinctions at the variable and person level in primary school-aged children. Aggressive Behavior 44, 5 (2018), 431–441.
[17]
M. Cristani, V. Murino, and A. Vinciarelli. 2010. Socially intelligent surveillance and monitoring: Analysing social dimensions of physical space. In 2010 Conf. on Computer Vision and Pattern Recognition - Workshops. IEEE Computer Society, NY, USA, 51–58. https://doi.org/10.1109/CVPRW.2010.5543179
[18]
N. Cummins, J. Joshi, A. Dhall, V.n Sethu, R. Goecke, and J. Epps. 2013. Diagnosis of depression by behavioural signals: a multimodal approach. In Proc. of the 3rd ACM Int. workshop on Audio/visual emotion challenge. ACM, NY, USA, 11–20.
[19]
O.J. Dunn. 1964. Multiple Comparisons Using Rank Sums. Technometrics 6, 3 (1964), 241–252.
[20]
J. O. Egede, S. Song, T. A Olugbade, C. Wang, C De C Amanda, H.g Meng, M. Aung, N. D. Lane, M. Valstar, and N. Bianchi-Berthouze. 2020. Emopain challenge 2020: Multimodal pain evaluation from facial and bodily expressions. In 2020 15th IEEE Int. Conf. on Automatic Face and Gesture Recognition. IEEE Computer Society, NY, USA, 849–856.
[21]
R. el Kaliouby and P. Robinson. 2005. Real-time inference of complex mental states from facial expressions and head gestures. In Real-Time Vision for Human-Computer Interaction. Springer, Boston, MA, 181–200.
[22]
F. Renee G. Moraga, S. Klein Tuente, S. Perrin, P. Enebrink, K. Sygel, W. Veling, and M. Wallinius. 2022. New developments in virtual reality-assisted treatment of aggression in forensic settings: The case of VRAPT. Frontiers in Virtual Reality 2 (2022), 174.
[23]
G. Giannakakis, D. Manousos, P. Simos, and M. Tsiknakis. 2018. Head movements in context of speech during stress induction. In Proc. of Int. Conf. on Automatic Face & Gesture Recognition. IEEE Computer Society, NY, USA, 710–714.
[24]
H. Gunes and M. Pantic. 2010. Dimensional emotion prediction from spontaneous head gestures for interaction with sensitive artificial listeners. In Proc. of Intelligent Virtual Agents. Springer, Boston, MA, 371–377.
[25]
H. Gunes and M. Piccardi. 2007. Bi-modal emotion recognition from expressive face and body gestures. J. of Network and Computer Applications 30, 4 (2007), 1334–1345.
[26]
U. Hadar, T.J. Stiener, E.C. Grant, and F. C. Rose. 1983. Head movement correlates of juncture and stress at sentence level. Language and Speech 26 (1983), 117–129.
[27]
U. Hadar, T.J. Stiener, E.C. Grant, and F. C. Rose. 1983. Kinematics of head movements accompanying speech during conversation. Human Movement Science 2 (1983), 35–45.
[28]
Z. Hammal, J.F. Cohn, and D.T. George. 2014. Interpersonal Coordination of Head Motion in Distressed Couples. IEEE Trans. on Affective Computing 5, 2 (2014), 155–167.
[29]
Z. Hammal, J.F. Cohn, C.L. Heike, and M.L. Speltz. 2015. Automatic Measurement of head and Facial Movement for analysis and Detection of infants’ Positive and negative affect. Frontier ICT 2, 21 (2015), 11 pages.
[30]
Z. Hammal, J.F. Cohn, C. Heike, and M.L. Speltz. 2015. What can head and facial movements convey about positive and negative affect?. In 2015 Int. Conf. on Affective Computing and Intelligent Interaction (ACII). IEEE Computer Society, NY, USA, 281–287.
[31]
J.A. Harrigan. 1985. Listeners’ Body Movements and Speaking Turns. Communication Research 12 (1985), 233––250.
[32]
S. Duncan Jr.1972. Some signals and rules for taking speaking turns in conversations. J. of Personality and Social Psychology 23, 2 (1972), 283–292.
[33]
S. Duncan Jr. and D.W. Fiske. 1977. Face-to-face interaction: Research, methods, and theory. L. Erlbaum Associates, New Jersey.
[34]
A. Kendon. 1967. Some functions of gaze-direction in social interaction. Acta Psychologica 26 (1967), 22–63.
[35]
Samuel Kim, Fabio Valente, Maurizio Filippone, and Alessandro Vinciarelli. 2014. Predicting Continuous Conflict Perceptionwith Bayesian Gaussian Processes. IEEE Trans. on Affective Computing 5, 2 (2014), 187–200. https://doi.org/10.1109/TAFFC.2014.2324564
[36]
M.L. Knapp and J.A. Hall. 2010. Nonverbal Behavior in Human Communication (7 ed.). Wadsworth, Cengage Learning, Boston, MA.
[37]
K. Krippendorff. 2007. Computing Krippendorff’s alpha reliability., 43 pages. Departmental papers (ASC).
[38]
W.H. Kruskal and W.A. Wallis. 1952. Use of Ranks in One-Criterion Variance Analysis. J. of the American Statistical Association 47, 260 (1952), 583–621.
[39]
J. Lafferty, A. McCallum, and F. Pereira. 2001. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data. In Proc. 18th Int. Conf. on Machine Learning. Morgan Kaufmann, San Francisco, CA, 282–289.
[40]
Iulia Lefter, Alice Baird, Lukas Stappen, and Björn W Schuller. 2022. A cross-corpus speech-based analysis of escalating negative interactions. Frontiers is Computer Science, Human Media Interaction 4 (2022), 749804.
[41]
I. Lefter, G. Burghouts, and L. Rothkrantz. 2015. Recognizing stress using semantics and modulation of speech and gestures.IEEE Trans. on Affective Computing 7, 2 (2015), 162–175.
[42]
I. Lefter and S. Fitrianie. 2018. The Multimodal Dataset of Negative Affect and Aggression: A Validation Study. In Proc. of the 2018 on Int. Conf. on Multimodal Interaction. ACM, NY, USA, 376–383.
[43]
Iulia Lefter and Siska Fitrianie. 2023. On Head Motion for Recognizing Aggression and Negative Affect during Speaking and Listening: Supplementary Materials. (2023). https://doi.org/10.4121/0478556b-ff3a-40ac-9662-f0e0a18ab4ee
[44]
I. Lefter and C.M. Jonker. 2017. Aggression recognition using overlapping speech. In 2017 7th Int. Conf. on Affective Computing and Intelligent Interaction (ACII). IEEE Computer Society, NY, USA, 299–304.
[45]
I. Lefter, C.M. Jonker, S. Klein Tuente, W. Veling, and S. Bogaerts. 2017. NAA: A multimodal database of negative affect and aggression. In 2017 7th Int. Conf. on Affective Computing and Intelligent Interaction. IEEE Computer Society, NY, USA, 21–27.
[46]
S. Madan, M. Gahalawat, T. Guha, and R. Subramanian. 2021. Head Matters: Explainable Human-centered Trait Prediction from Head Motion Dynamics. In Proc. of the 2021 Int. Conf. on Multimodal Interaction. ACM, NY, USA, 435–443.
[47]
E. Z. McClave. 2000. Linguistic functions of head movements in the context of speech. J. of Pragmatics 32, 7 (2000), 855 – 878.
[48]
Nathan Moore, Naseem Ahmadpour, Martin Brown, Philip Poronnik, and Jennifer Davids. 2022. Designing Virtual Reality–Based Conversational Agents to Train Clinicians in Verbal De-escalation Skills: Exploratory Usability Study. JMIR Serious Games 10, 3 (2022), e38669.
[49]
E. Murphy-Chutorian and M.M. Trivedi. 2009. Head Pose Estimation in Computer Vision: A Survey. IEEE Trans. on Pattern Analysis and Machine Intelligence 31, 4 (2009), 607–626.
[50]
S. Nayak, B. Nagesh, A. Routray, and M. Sarma. 2021. A Human–Computer Interaction framework for emotion recognition through time-series thermal video sequences. Computers & Electrical Engineering 93 (2021), 107280.
[51]
F. Noroozi, D. Kaminska, C. Corneanu, T. Sapinski, S. Escalera, and G. Anbarjafari. 2018. Survey on Emotional Body Gesture Recognition. IEEE Trans. on Affective Computing 12 (2018), 505–523.
[52]
T. Parr, K. Turgutlu, C. Csiszar, and J. Howard. 2018. Beware Default Random Forest Importances. https://explained.ai/rf-importance/.
[53]
A. Pease and B. Pease. 2017. The definitive book of body language. Orion, London, England.
[54]
Bruno Peixoto, Bahram Lavi, João Paulo Pereira Martin, Sandra Avila, Zanoni Dias, and Anderson Rocha. 2019. Toward subjective violence detection in videos. In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE Computer Society, NY, USA, 8276–8280.
[55]
Tim Polzehl, Alexander Schmitt, Florian Metze, and Michael Wagner. 2011. Anger recognition in speech using acoustic and linguistic cues. Speech Communication 53, 9-10 (2011), 1198–1209.
[56]
P. Rani, N.n Sarkar, and J. Adams. 2007. Anxiety-based affective communication for implicit human–machine interaction. Advanced Engineering Informatics 21, 3 (2007), 323–334.
[57]
A. Samanta and T. Guha. 2017. On the role of head motion in affective expression. In 2017 IEEE Int. Conf. on Acoustics, Speech and Signal Processing. IEEE Computer Society, NY, USA, 2886–2890.
[58]
K. R. Scherer and T. Bänziger. 2010. On the use of actor portrayals in research on emotional expression. In Blueprint for affective computing: A sourcebook, K. R. Scherer, T. Bänziger, and E. B. Roesch (Eds.). Oxford university Press, Oxford, England, 166–176.
[59]
R. De Silva and N. Bianchi-Berthouze. 2004. Modeling human affective postures: An information theoretic characterization of posture features. J. Computer Animation and Virtual Worlds 15, 3/4 (2004), 269–276.
[60]
H. J. Smith and M. Neff. 2018. Communication Behavior in Embodied Virtual Reality. In Proc. of the 2018 CHI Conf. on Human Factors in Computing Systems (Montreal QC, Canada) (CHI ’18). ACM, NY, USA, 1–12.
[61]
Meishu Song, Adria Mallol-Ragolta, Emilia Parada-Cabaleiro, Zijiang Yang, Shuo Liu, Zhao Ren, Ziping Zhao, and Björn W Schuller. 2021. Frustration recognition from speech during game interaction using wide residual networks. Virtual Reality & Intelligent Hardware 3, 1 (2021), 76–86.
[62]
Stephanie Tan, David MJ Tax, and Hayley Hung. 2021. Multimodal Joint Head Orientation Estimation in Interacting Groups via Proxemics and Interaction Dynamics. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, 1 (2021), 1–22.
[63]
B.-J. Theobald, I. Matthews, M. Mangini, J.R. Spies, T. Brick, J.F. Cohn, and S.M. Boker. 2009. Mapping and manipulating facial expression. Language and Speech 52, 2/3 (2009), 369–386.
[64]
Theodoros Theodoridis and Huosheng Hu. 2013. Modeling Aggressive Behaviors With Evolutionary Taxonomers. IEEE Trans. on Human-Machine Systems 43, 3 (2013), 302–313. Conf. Name: IEEE Trans. on Human-Machine Systems.
[65]
P. Werner, A. Al-Hamadi, K. Limbrecht-Ecklundt, S. Walter, and H.C. Traue. 2018. Head movements and postures as pain behavior. PLoS ONE 13, 2 (2018), e0192767.
[66]
T. Xue, A. E. Ali, G. Ding, and P. Cesar. 2021. Investigating the relationship between momentary emotion self-reports and head and eye movements in hmd-based 360 vr video watching. In Extended Abstracts of the 2021 CHI Conf. on Human Factors in Computing Systems. ACM, NY, USA, 1–8.
[67]
V. Yngve. 1970. On getting a word in edgewise. Papers from the sixth Regional Meeting of the Chicago Linguistic Society 6-7 (1970), 567–578.
[68]
Wojtek Zajdel, Johannes D Krijnders, Tjeerd Andringa, and Dariu M Gavrila. 2007. CASSANDRA: audio-video sensor fusion for aggression detection. In 2007 IEEE conference on advanced video and signal based surveillance. IEEE Computer Society, NY, USA, 200–205.
[69]
S. Zhao, S. Wang, M. Soleymani, D. Joshi, and Q. Ji. 2019. Affective computing for large-scale heterogeneous multimedia data: A survey. ACM Trans. on Multimedia Computing, Communications, and Applications (TOMM) 15, 3s (2019), 1–32.

Cited By

View all
  • (2024)FacePsy: An Open-Source Affective Mobile Sensing System - Analyzing Facial Behavior and Head Gesture for Depression Detection in Naturalistic SettingsProceedings of the ACM on Human-Computer Interaction10.1145/36765058:MHCI(1-32)Online publication date: 24-Sep-2024

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
ICMI '23: Proceedings of the 25th International Conference on Multimodal Interaction
October 2023
858 pages
ISBN:9798400700552
DOI:10.1145/3577190
This work is licensed under a Creative Commons Attribution-NonCommercial International 4.0 License.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 09 October 2023

Check for updates

Author Tags

  1. Affective Aggression
  2. Affective Computing
  3. Emotion
  4. Head Motion
  5. NAA database
  6. Speaking and Listening Behavior.

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICMI '23
Sponsor:

Acceptance Rates

Overall Acceptance Rate 453 of 1,080 submissions, 42%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)182
  • Downloads (Last 6 weeks)27
Reflects downloads up to 18 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)FacePsy: An Open-Source Affective Mobile Sensing System - Analyzing Facial Behavior and Head Gesture for Depression Detection in Naturalistic SettingsProceedings of the ACM on Human-Computer Interaction10.1145/36765058:MHCI(1-32)Online publication date: 24-Sep-2024

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Login options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media