Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3614008.3614062acmotherconferencesArticle/Chapter ViewAbstractPublication PagesspmlConference Proceedingsconference-collections
research-article

Automated Speech Recognition System for Dispatching Call Recordings in The Underground Coal Mines

Published: 17 October 2023 Publication History

Abstract

In this paper, we proposed an automated speech recognition system focus on the dispatching call recordings in the underground coal mines which promoted the development of intelligent coal mining. The main challenges of the speech recognition system are the noise of recordings and the dialect speech. We employed a voice activity detection module to preprocess the recordings, this module is able to reduce the noise and segment the long recording speech; then the Conformer model with CTC algorithm is utilized to train the ASR module. To get better performance, the WenetSpeech pretrained model is embedded for fine-tuning. The result shows that compared with the other general speech recognition systems, our ASR system has great advance in recognizing the dispatching call recordings of Huaibei dialect.

References

[1]
Sohn, Jongseo, Nam Soo Kim, and Wonyong Sung. "A statistical model-based voice activity detection." IEEE signal processing letters 6.1 (1999): 1-3.
[2]
Gulati, Anmol, "Conformer: Convolution-augmented transformer for speech recognition." arXiv preprint arXiv:2005.08100 (2020).
[3]
Sharma, Mayank, "A comprehensive empirical review of modern voice activity detection approaches for movies and TV shows." Neurocomputing (2022).
[4]
D. Ying, Y. Yan, J. Dang and F. K. Soong, "Voice Activity Detection Based on an Unsupervised Learning Framework," in IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, no. 8, pp. 2624-2633, Nov. 2011.
[5]
Woo, Kyoung-Ho, "Robust voice activity detection algorithm for estimating noise spectrum." Electronics Letters 36.2 (2000): 180-181.
[6]
J. Ramírez and J. C. Segura, “Efficient voice activity detection algorithms using long-term speech information,” Speech Commun., vol. 42, no. 3, pp. 271–287, 2004.
[7]
Greff, Klaus, "LSTM: A search space odyssey." IEEE transactions on neural networks and learning systems 28.10 (2016): 2222-2232.
[8]
Ferroni, Giacomo, "A deep neural network approach for voice activity detection in multi-room domestic scenarios." 2015 International Joint Conference on Neural Networks (IJCNN). IEEE, 2015.
[9]
K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, arXiv preprint arXiv:1409.1556.
[10]
Davis KH, Biddulph R, Balashek S (1952) Automatic recognition of spoken digits. J Acoust Soc Am 24(6):637–642
[11]
Juang, B‐H. "On the hidden Markov model and dynamic time warping for speech recognition—A unified view." AT&T Bell Laboratories Technical Journal 63.7 (1984): 1213-1243.
[12]
Gales, Mark JF. "Maximum likelihood linear transformations for HMM-based speech recognition." Computer speech & language 12.2 (1998): 75-98.
[13]
Li, Jinyu, "Improving wideband speech recognition using mixed-bandwidth training data in CD-DNN-HMM." 2012 IEEE Spoken Language Technology Workshop (SLT). IEEE, 2012.
[14]
Soltau, Hagen, Hank Liao, and Hasim Sak. "Neural speech recognizer: Acoustic-to-word LSTM model for large vocabulary speech recognition." arXiv preprint arXiv:1610.09975 (2016).
[15]
Mamiya, Yoshitaka, "Lightly supervised GMM VAD to use audiobook for speech synthesiser." 2013 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 2013.
[16]
Zhang, B., Lv, H., Guo, P., Shao, Q., Yang, C., Xie, L., Xu, X., Bu, H., Chen, X., Zeng, C., Wu, D., & Peng, Z. (2021). WenetSpeech: A 10000+ Hours Multi-domain Mandarin Corpus for Speech Recognition. ArXiv. /abs/2110.03370.
[17]
Klakow, Dietrich, and Jochen Peters. "Testing the correlation of word error rate and perplexity." Speech Communication 38.1-2 (2002): 19-28.

Index Terms

  1. Automated Speech Recognition System for Dispatching Call Recordings in The Underground Coal Mines

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    SPML '23: Proceedings of the 2023 6th International Conference on Signal Processing and Machine Learning
    July 2023
    383 pages
    ISBN:9798400707575
    DOI:10.1145/3614008
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 17 October 2023

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Funding Sources

    • Fundamental Research Funds for the Central Universities

    Conference

    SPML 2023

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 25
      Total Downloads
    • Downloads (Last 12 months)20
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 16 Nov 2024

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media