Nothing Special   »   [go: up one dir, main page]

Skip to main content

Operating a Robot by Nonverbal Voice Expressed with Acoustic Features

  • Conference paper
  • First Online:
Intelligent Autonomous Systems 15 (IAS 2018)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 867))

Included in the following conference series:

Abstract

This paper proposes methods for operating a robot by nonverbal voice. These methods enable operators to operate multi-degrees of freedom simultaneously and operate a robot intuitively by nonverbal voice by associating the nonverbal voice, tongue position and the coordinate of the robot’s hand. The voice is defined by formants or Mel Frequency Cepstral Coefficients (MFCC). Formants and MFCC are acoustic features and they show the characteristics of the vocal tract such as the mouth and the tongue. We propose two methods. One is the method in which voice expressed with overlapped formants ranges are used to change variable about robots’ operation. This method enables operators to operate multi-degrees of freedom simultaneously by nonverbal voice. The other is the method that operators tongue positions are distinguished by nonverbal voice. These tongue positions correspond to the coordinate of the robot’s hand and it enables the operators to operate a robot intuitively. We found the feasibility of the methods through experiments of simple tasks. These methods can realize operating a robot intuitively in continuous values by voice and can be utilized for user-friendly system.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Grondin, F., Michaud, F.: Robust speech/non-speech discrimination based on pitch estimation for mobile robots. In: 2016 IEEE International Conference on Robotics and Automation (ICRA), pp. 1650–1655, May 2016

    Google Scholar 

  2. Harada, S., Saponas, T.S., Landay, J.A.: Voicepen: augmenting pen input with simultaneous non-linguistic vocalization. In: the 9th Conference on Multimodal Interfaces, pp. 178–185 (2007)

    Google Scholar 

  3. Harada, S., Wobbrock, J.O., Landay, J.A.: Voicedraw: a hands-free voice-driven drawing application for people with motor impairments. In: The 9th Conference on Computers and Accessibility, pp. 27–34 (2007)

    Google Scholar 

  4. House, B., Malkin, J., Bilmes, J.: The voicebot: a voice controlled robot arm. In: The Conference on Human Factors in Computing Systems, pp. 183–192 (2009)

    Google Scholar 

  5. Igarashi, T., Hughes, J.F.: Voice as sound: using non-verbal voice input for interactive control. In: The 14th Symposium on User Interface Software and Technology, pp. 155–156 (2001)

    Google Scholar 

  6. Malkin, J., Li, X., Harada, S., Landay, J., Bilmes, J.: The vocal joystick engine v1.0. computational speech. Language 25(3), 535–555 (2011)

    Google Scholar 

  7. Mihara, Y., Shibayama, E., Takahashi, S.: The migratory cursor: accurate speech-based cursor movement by moving multiple ghost cursors using non-verbal vocalizations. In: The 7th Conference on Computers and Accessibility. Assets 2005, pp. 76–83 (2005)

    Google Scholar 

  8. Mizuuchi, I., Fujimoto, J., Sodeyama, Y., Yamamoto, K., Okada, K., Inaba, M.: A kitchen assistant manipulation system of a variety of dishes based on shape estimation with tracing dish surfaces by sensing proximity and touch information. J. Robot. Soc. Jpn. 30(9), 889–898 (2012)

    Article  Google Scholar 

  9. Nakadai, K., Okuno, H.G., Nakajima, H., Hasegawa, Y., Tsujino, H.: An open source software system for robot audition hark and its evaluation. In: Humanoids 2008 - 8th IEEE-RAS International Conference on Humanoid Robots, pp. 561–566, December 2008

    Google Scholar 

  10. Sakamoto, D., Komatsu, T., Igarashi, T.: Voice augmented manipulation: using paralinguistic information to manipulate mobile devices. In: The 15th Conference on Human-Computer Interaction with Mobile Devices and Services, pp. 69–78 (2013)

    Google Scholar 

  11. Sears, A., Young, M.: Physical disabilities and computing technologies: An Analysis of Impairments. In: The Human-Computer Interaction Handbook, pp. 482–503 (2003)

    Google Scholar 

  12. Sporka, A.J., Slavík, P.: Vocal control of a radio-controlled car. In: SIGACCESS, pp. 3–8 (2008)

    Google Scholar 

  13. Yoshida, T., Nakadai, K., Okuno, H.G.: Automatic speech recognition improved by two-layered audio-visual integration for robot audition. In: 2009 9th IEEE-RAS International Conference on Humanoid Robots, pp. 604–609, December 2009

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shizuka Takahashi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Takahashi, S., Mizuuchi, I. (2019). Operating a Robot by Nonverbal Voice Expressed with Acoustic Features. In: Strand, M., Dillmann, R., Menegatti, E., Ghidoni, S. (eds) Intelligent Autonomous Systems 15. IAS 2018. Advances in Intelligent Systems and Computing, vol 867. Springer, Cham. https://doi.org/10.1007/978-3-030-01370-7_45

Download citation

Publish with us

Policies and ethics