Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3536220.3563686acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
short-paper

Real-time Public Speaking Anxiety Prediction Model for Oral Presentations

Published: 07 November 2022 Publication History

Abstract

Oral presentation skills are essential for most people’s academic and career development. However, due to public speaking anxiety, many people find oral presentations challenging and often avoid them to the detriment of their careers. Public speaking anxiety interventions that help presenters manage their anxiety as it occurs during a presentation can help many presenters. In this paper, we present a model for assessing public speaking anxiety during a presentation—a first step towards developing real-time anxiety interventions. We present our method for ground truth data collection and the results of neural network models for real-time anxiety detection using audio data. Our results show that using an LSTM model we can predict moments of speaking anxiety during a presentation.

References

[1]
Deborah C Beidel and Samuel M Turner. 1998. Shy children, phobic adults: Nature and treatment of social phobia. American Psychological Association, England.
[2]
Melissa A Birkett. 2011. The Trier Social Stress Test protocol for inducing psychological stress. JoVE (Journal of Visualized Experiments) 56 (2011), e3238.
[3]
John B Bishop, Karen W Bauer, and Elizabeth Trezise Becker. 1998. A survey of counseling needs of male and female college students. Journal of College Student Development 39, 2 (1998), 205–210.
[4]
Graham D Bodie. 2010. A racing heart, rattling knees, and ruminative thoughts: Defining, explaining, and treating public speaking anxiety. Communication education 59, 1 (2010), 70–105.
[5]
François Chollet and others. 2018. Keras: The Python Deep Learning library. Astrophysics Source Code Library, record ascl:1806.022., ascl:1806.022 pages. ascl:1806.022
[6]
Katarina Dedovic, Robert Renwick, Najmeh Khalili Mahani, Veronika Engert, Sonia J Lupien, and Jens C Pruessner. 2005. The Montreal Imaging Stress Task: using functional imaging to investigate the effects of perceiving and processing psychosocial stress in the human brain. Journal of Psychiatry and Neuroscience 30, 5 (2005), 319–325.
[7]
Florian Eyben, Klaus R. Scherer, Björn W. Schuller, Johan Sundberg, Elisabeth André, Carlos Busso, Laurence Y. Devillers, Julien Epps, Petri Laukka, Shrikanth S. Narayanan, and Khiet P. Truong. 2016. The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing. IEEE Transactions on Affective Computing 7, 2 (2016), 190–202. https://doi.org/10.1109/TAFFC.2015.2457417
[8]
Florian Eyben, Martin Wöllmer, and Björn Schuller. 2010. Opensmile: The Munich Versatile and Fast Open-Source Audio Feature Extractor. In Proceedings of the 18th ACM International Conference on Multimedia (Firenze, Italy) (MM ’10). Association for Computing Machinery, New York, NY, USA, 1459–1462. https://doi.org/10.1145/1873951.1874246
[9]
Yin Fan, Xiangju Lu, Dian Li, and Yuanliu Liu. 2016. Video-Based Emotion Recognition Using CNN-RNN and C3D Hybrid Networks. In Proceedings of the 18th ACM International Conference on Multimodal Interaction (Tokyo, Japan) (ICMI ’16). Association for Computing Machinery, New York, NY, USA, 445–450. https://doi.org/10.1145/2993148.2997632
[10]
Kexin Feng, Megha Yadav, Md Nazmus Sakib, Amir Behzadan, and Theodora Chaspari. 2019. Estimating Public Speaking Anxiety from Speech Signals Using Unsupervised Transfer Learning. In 2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP). IEEE, Ottawa, ON, Canada, 1–5. https://doi.org/10.1109/GlobalSIP45357.2019.8969502
[11]
Raul Fernandez. 2004. A computational model for the automatic recognition of affect in speech. Ph. D. Dissertation. Massachusetts Institute of Technology.
[12]
Barbara F Fuller, Yoshiyuki Horii, and Douglas A Conner. 1992. Validity and reliability of nonverbal voice measures as indicators of stressor-provoked anxiety. Research in nursing & health 15, 5 (1992), 379–389.
[13]
Jeffrey Girard. 2014. CARMA: Software for Continuous Affect Rating and Media Annotation. Journal of Open Research Software 2 (2014). https://doi.org/10.5334/jors.ar
[14]
Alexander M Goberman, Stephanie Hughes, and Todd Haydock. 2011. Acoustic characteristics of public speaking: Anxiety and practice effects. Speech communication 53, 6 (2011), 867–876.
[15]
Muriel A Hagenaars and Agnes van Minnen. 2005. The effect of fear on paralinguistic aspects of speech in patients with panic disorder with agoraphobia. Journal of Anxiety Disorders 19, 5 (2005), 521–537.
[16]
Jinni A Harrigan, Kelly M Harrigan, Beverly A Sale, and Robert Rosenthal. 1996. Detecting anxiety and defensiveness from visual and auditory cues. Journal of Personality 64, 3 (1996), 675–709.
[17]
Karen Hovsepian, Mustafa al’Absi, Emre Ertin, Thomas Kamarck, Motohiro Nakajima, and Santosh Kumar. 2015. CStress: Towards a Gold Standard for Continuous Stress Assessment in the Mobile Environment. In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing (Osaka, Japan) (UbiComp ’15). Association for Computing Machinery, New York, NY, USA, 493–504. https://doi.org/10.1145/2750858.2807526
[18]
Ping Hu, Dongqi Cai, Shandong Wang, Anbang Yao, and Yurong Chen. 2017. Learning Supervised Scoring Ensemble for Emotion Recognition in the Wild. In Proceedings of the 19th ACM International Conference on Multimodal Interaction (Glasgow, UK) (ICMI ’17). Association for Computing Machinery, New York, NY, USA, 553–560. https://doi.org/10.1145/3136755.3143009
[19]
Tom Johnstone, Carien M Van Reekum, Tanja Bänziger, Kathryn Hird, Kim Kirsner, and Klaus R Scherer. 2007. The effects of difficulty and gain versus loss on vocal physiology and acoustics. Psychophysiology 44, 5 (2007), 827–837.
[20]
Everlyne Kimani, Timothy Bickmore, Ha Trinh, and Paola Pedrelli. 2019. You’ll be Great: Virtual Agent-based Cognitive Restructuring to Reduce Public Speaking Anxiety. In 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII). IEEE, Cambridge, UK, 641–647.
[21]
Everlyne Kimani, Ameneh Shamekhi, and Timothy Bickmore. 2021. Just breathe: Towards real-time intervention for public speaking anxiety. Smart Health 19(2021), 100146. https://doi.org/10.1016/j.smhl.2020.100146
[22]
Diederik P. Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimization. https://doi.org/10.48550/ARXIV.1412.6980
[23]
Petri Laukka, Clas Linnman, Fredrik Åhs, Anna Pissiota, Örjan Frans, Vanda Faria, Åsa Michelgård, Lieuwe Appel, Mats Fredrikson, and Tomas Furmark. 2008. In a nervous voice: Acoustic analysis and perception of anxiety in social phobics’ speech. Journal of Nonverbal Behavior 32, 4 (2008), 195–214.
[24]
Lan Li, Fei Yu, Dongquan Shi, Jianping Shi, Zongjun Tian, Jiquan Yang, Xingsong Wang, and Qing Jiang. 2017. Application of virtual reality technology in clinical medicine. American journal of translational research 9, 9 (2017), 3867.
[25]
Xi Li, Jidong Tao, Michael T Johnson, Joseph Soltis, Anne Savage, Kirsten M Leong, and John D Newman. 2007. Stress and emotion classification using jitter and shimmer features. In 2007 IEEE International Conference on Acoustics, Speech and Signal Processing-ICASSP’07, Vol. 4. IEEE, Honolulu, HI, USA, IV–1081.
[26]
Phil Lopes, Georgios N Yannakakis, and Antonios Liapis. 2017. Ranktrace: Relative and unbounded affect annotation. In 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII). IEEE, San Antonio, TX, USA, 158–163.
[27]
Konstantinos Makantasis, Antonios Liapis, and Georgios N Yannakakis. 2021. The pixels and sounds of emotion: General-purpose representations of arousal in games. IEEE Transactions on Affective Computing 12 (2021), 1–1.
[28]
James C. McCroskey and Linda L. McCroskey. 1988. Self‐report as an approach to measuring communication competence. Communication Research Reports 5, 2 (1988), 108–113. https://doi.org/10.1080/08824098809359810
[29]
Ehsanul Haque Nirjhar, Amir Behzadan, and Theodora Chaspari. 2020. Exploring bio-behavioral signal trajectories of state anxiety during public speaking. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, Barcelona, Spain, 1294–1298.
[30]
Ehsanul Haque Nirjhar, Amir H. Behzadan, and Theodora Chaspari. 2021. Knowledge- and Data-Driven Models of Multimodal Trajectories of Public Speaking Anxiety in Real and Virtual Settings. In Proceedings of the 2021 International Conference on Multimodal Interaction (Montréal, QC, Canada) (ICMI ’21). Association for Computing Machinery, New York, NY, USA, 712–716. https://doi.org/10.1145/3462244.3479964
[31]
Gordon L Paul. 1966. Insight vs. desensitization in psychotherapy: An experiment in anxiety reduction. Stanford University Press, California, USA. 402–403 pages.
[32]
Silke Paulmann, Desire Furnes, Anne Ming Bøkenes, and Philip J Cozzolino. 2016. How psychological stress affects emotional prosody. Plos one 11, 11 (2016), e0165022.
[33]
Rosalind W Picard. 2000. Affective computing. MIT press, London, UK.
[34]
Katarzyna Pisanski and Piotr Sorokowski. 2021. Human stress detection: cortisol levels in stressed speakers predict voice-based judgments of stress. Perception 50, 1 (2021), 80–87.
[35]
Kurt Plarre, Andrew Raij, Syed Monowar Hossain, Amin Ahsan Ali, Motohiro Nakajima, Mustafa Al’Absi, Emre Ertin, Thomas Kamarck, Santosh Kumar, Marcia Scott, 2011. Continuous inference of psychological stress from sensory measurements collected in the natural environment. In Proceedings of the 10th ACM/IEEE international conference on information processing in sensor networks. IEEE, Chicago, IL, USA, 97–108.
[36]
Alec Pollard and Gibson Henderson. 1988. Four types of social phobia in a community sample. Journal of Nervous and Mental Disease 176, 7 (1988).
[37]
Melissa J Ree, Davina French, Colin MacLeod, and Vance Locke. 2008. Distinguishing cognitive and somatic dimensions of state and trait anxiety: Development and validation of the State-Trait Inventory for Cognitive and Somatic Anxiety (STICSA). Behavioural and Cognitive Psychotherapy 36, 3 (2008), 313–332.
[38]
Patrice Renaud and Jean-Pierre Blondin. 1997. The stress of Stroop performance: Physiological and emotional responses to color–word interference, task pacing, and pacing speed. International Journal of Psychophysiology 27, 2 (1997), 87–97.
[39]
Klaus R Scherer. 1986. Vocal affect expression: a review and a model for future research.Psychological bulletin 99, 2 (1986), 143.
[40]
Klaus R Scherer and Agnes Moors. 2019. The emotion process: Event appraisal and component differentiation. Annual review of psychology 70 (2019), 719–745.
[41]
Björn Schuller, Anton Batliner, Stefan Steidl, and Dino Seppi. 2011. Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge. Speech communication 53, 9-10 (2011), 1062–1087.
[42]
Sadat Shahriar and Yelin Kim. 2019. Audio-Visual Emotion Forecasting: Characterizing and Predicting Future Emotion Using Deep Learning. In 2019 14th IEEE International Conference on Automatic Face Gesture Recognition (FG 2019). IEEE, Lille, France, 1–7. https://doi.org/10.1109/FG.2019.8756599
[43]
George M Slavich, Sara Taylor, and Rosalind W Picard. 2019. Stress measurement using speech: Recent advancements, validation issues, and ethical and privacy considerations. Stress 22, 4 (2019), 408–413.
[44]
Hariharan Subramonyam, Colleen Seifert, Priti Shah, and Eytan Adar. 2020. TexSketch: Active Diagramming through Pen-and-Ink Annotations. Association for Computing Machinery, New York, NY, USA, 1–13. https://doi.org/10.1145/3313831.3376155
[45]
Barry Tanner. 2012. Validity of Global Physical and Emotional SUDS. Applied psychophysiology and biofeedback 37 (03 2012), 31–4. https://doi.org/10.1007/s10484-011-9174-x
[46]
Xiao Yang, Nilam Ram, Jessica P Lougheed, Peter Molenaar, and Tom Hollenstein. 2019. Adolescents’ emotion system dynamics: Network-based analysis of physiological and emotional experience.Developmental psychology 55, 9 (2019), 1982.
[47]
Zhaoyan Zhang. 2016. Mechanics of human voice production and control. The journal of the acoustical society of america 140, 4 (2016), 2614–2635.
[48]
Zoom. 2021. Video Conferencing, Web Conferencing, Webinars, Screen Sharing. https://zoom.us/. Accessed: 2021-01-11.

Cited By

View all
  • (2023)Sharing Speaker Heart Rate with the Audience Elicits Empathy and Increases PersuasionPersuasive Technology10.1007/978-3-031-30933-5_1(3-21)Online publication date: 19-Apr-2023

Index Terms

  1. Real-time Public Speaking Anxiety Prediction Model for Oral Presentations

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      ICMI '22 Companion: Companion Publication of the 2022 International Conference on Multimodal Interaction
      November 2022
      225 pages
      ISBN:9781450393898
      DOI:10.1145/3536220
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 07 November 2022

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. Affective computing
      2. public speaking anxiety
      3. real-time prediction
      4. speech

      Qualifiers

      • Short-paper
      • Research
      • Refereed limited

      Conference

      ICMI '22
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 453 of 1,080 submissions, 42%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)80
      • Downloads (Last 6 weeks)12
      Reflects downloads up to 21 Nov 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2023)Sharing Speaker Heart Rate with the Audience Elicits Empathy and Increases PersuasionPersuasive Technology10.1007/978-3-031-30933-5_1(3-21)Online publication date: 19-Apr-2023

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format.

      HTML Format

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media