default search action
Computer Speech & Language, Volume 46
Volume 46, November 2017
- Jindrich Matousek, Daniel Tihelka:
Anomaly-based annotation error detection in speech-synthesis corpora. 1-35 - Carmen Magariños, Paula Lopez-Otero, Laura Docío Fernández, Eduardo Rodríguez Banga, Daniel Erro, Carmen García-Mateo:
Reversible speaker de-identification using pre-trained transformation functions. 36-52 - Hossein Zeinali, Hossein Sameti, Lukás Burget, Jan Cernocký:
Text-dependent speaker verification based on i-vectors, Neural Networks and Hidden Markov Models. 53-71 - Brecht Desplanques, Kris Demuynck, Jean-Pierre Martens:
Adaptive speaker diarization of broadcast news based on factor analysis. 72-93 - Janez Starc, Dunja Mladenic:
Constructing a Natural Language Inference dataset using generative neural networks. 94-112 - Scott Piao, Fraser Dallachy, Alistair Baron, Jane Demmen, Steve Wattam, Philip Durkin, James McCracken, Paul Rayson, Marc Alexander:
A time-sensitive historical thesaurus-based semantic tagger for deep semantic annotation. 113-135 - Rachel G. Anushiya, P. Vijayalakshmi, T. Nagarajan:
Estimation of glottal closure instants from degraded speech using a phase-difference-based algorithm. 136-153 - Herman Kamper, Aren Jansen, Sharon Goldwater:
A segmental framework for fully-unsupervised large-vocabulary speech recognition. 154-174 - Christoph Draxler, Jonathan Harrington, Florian Schiel:
Towards the next generation of speech tools and corpora. 175-178 - Michael Pucher, Bettina Zillinger, Markus Toman, Dietmar Schabus, Cassia Valentini-Botinhao, Junichi Yamagishi, Erich Schmid, Thomas Woltron:
Influence of speaker familiarity on blind and visually impaired children's and young adults' perception of synthetic voices. 179-195 - Milos Cernak, Juan Rafael Orozco-Arroyave, Frank Rudzicz, Heidi Christensen, Juan Camilo Vásquez-Correa, Elmar Nöth:
Characterisation of voice quality of Parkinson's disease using differential phonological posterior features. 196-208 - Taehwan Kim, Jonathan Keane, Weiran Wang, Hao Tang, Jason Riggle, Gregory Shakhnarovich, Diane Brentari, Karen Livescu:
Lexicon-free fingerspelling recognition from video: Data, models, and signer adaptation. 209-232 - Shakti P. Rath:
Scalable algorithms for unsupervised clustering of acoustic data for speech recognition. 233-248 - Milica Gasic, Dilek Hakkani-Tür, Asli Celikyilmaz:
Spoken language understanding and interaction: machine learning for human-like conversational systems. 249-251 - Radek Fér, Pavel Matejka, Frantisek Grézl, Oldrich Plchot, Karel Veselý, Jan Honza Cernocký:
Multilingually trained bottleneck features in spoken language recognition. 252-267 - Heysem Kaya, Albert Ali Salah, Alexey Karpov, Olga V. Frolova, Aleksey Grigorev, Elena E. Lyakso:
Emotion, age, and gender classification in children's speech by humans and machines. 268-283 - Ingrid Zukerman, Andisheh Partovi:
Improving the understanding of spoken referring expressions through syntactic-semantic and contextual-phonetic error-correction. 284-310 - Young-Bum Kim, Karl Stratos, Ruhi Sarikaya:
A Framework for pre-training hidden-unit conditional random fields and its extension to long short term memory networks. 311-326 - Raymond W. M. Ng, Mauro Nicolao, Thomas Hain:
Unsupervised crosslingual adaptation of tokenisers for spoken language recognition. 327-342 - Harishchandra Dubey, Abhijeet Sangwan, John H. L. Hansen:
Using speech technology for quantifying behavioral characteristics in peer-led team learning sessions. 343-366 - Marta R. Costa-jussà, Alexandre Allauzen, Loïc Barrault, Kyunghyun Cho, Holger Schwenk:
Introduction to the special issue on deep learning approaches for machine translation. 367-373 - Jahn Heymann, Lukas Drude, Reinhold Haeb-Umbach:
A generic neural acoustic beamforming architecture for robust multi-channel speech processing. 374-385 - Jon Barker, Ricard Marxer, Emmanuel Vincent, Shinji Watanabe:
Multi-microphone speech recognition in everyday environments. 386-387 - Hendrik Barfuss, Christian Huemmer, Andreas Schwarz, Walter Kellermann:
Robust coherence-based spectral enhancement for speech recognition in adverse real-world environments. 388-400 - Takaaki Hori, Zhuo Chen, Hakan Erdogan, John R. Hershey, Jonathan Le Roux, Vikramjit Mitra, Shinji Watanabe:
Multi-microphone speech recognition integrating beamforming, robust feature extraction, and advanced DNN/RNN backend. 401-418 - Isidoros Rodomagoulakis, Athanasios Katsamanis, Gerasimos Potamianos, Panagiotis Giannoulis, Antigoni Tsiami, Petros Maragos:
Room-localized spoken command recognition in multi-room, multi-microphone environments. 419-443 - Sunit Sivasankaran, Emmanuel Vincent, Irina Illina:
A combined evaluation of established and new approaches for speech recognition in varied reverberation conditions. 444-460 - Ryu Takeda, Kazuhiro Nakadai, Kazunori Komatani:
Acoustic model training based on node-wise weight boundary model for fast and small-footprint deep neural networks. 461-480 - Payton Lin, Dau-Cheng Lyu, Fei Chen, Syu-Siang Wang, Yu Tsao:
Multi-style learning with denoising autoencoders for acoustic modeling in the internet of things (IoT). 481-495 - Ji-Won Cho, Jong-Hyeon Park, Joon-Hyuk Chang, Hyung-Min Park:
Bayesian feature enhancement using independent vector analysis and reverberation parameter re-estimation for noisy reverberant speech recognition. 496-516 - Yanhui Tu, Jun Du, Qing Wang, Xiao Bao, Li-Rong Dai, Chin-Hui Lee:
An information fusion framework with multi-channel feature concatenation and multi-perspective system combination for the deep-learning-based robust recognition of microphone array speech. 517-534 - Emmanuel Vincent, Shinji Watanabe, Aditya Arie Nugraha, Jon Barker, Ricard Marxer:
An analysis of environment, microphone and data simulation mismatches in robust speech recognition. 535-557 - Niko Moritz, Kamil Adiloglu, Jörn Anemüller, Stefan Goetze, Birger Kollmeier:
Multi-Channel Speech Enhancement and Amplitude Modulation Analysis for Noise Robust Automatic Speech Recognition. 558-573 - Alastair H. Moore, Pablo Peso Parada, Patrick A. Naylor:
Speech enhancement for robust automatic speech recognition: Evaluation using a baseline system and instrumental measures. 574-584 - Daniele Falavigna, Marco Matassoni, Shahab Jalalvand, Matteo Negri, Marco Turchi:
DNN adaptation by automatic quality estimation of ASR hypotheses. 585-604 - Jon Barker, Ricard Marxer, Emmanuel Vincent, Shinji Watanabe:
The third 'CHiME' speech separation and recognition challenge: Analysis and outcomes. 605-626
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.