default search action
SLT 2008: Goa, India
- Amitava Das, Srinivas Bangalore:
2008 IEEE Spoken Language Technology Workshop, SLT 2008, Goa, India, December 15-19, 2008. IEEE 2008, ISBN 978-1-4244-3472-5
Spoken Language Technology for Development
- Frederick Weber, Kalika Bali, Roni Rosenfeld, Kentaro Toyama:
Unexplored directions in spoken language technology for development. 1-4 - Amitava Das, Ohil K. Manyam, Makarand Tapaswi, Veeresh Taranalli:
Multilingual spoken-password based user authentication in emerging economies using cellular phone networks. 5-8 - Sebastian Spiegler, Bruno Golénia, Ksenia Shalonova, Peter A. Flach, Roger C. F. Tucker:
Learning the morphology of Zulu with different degrees of supervision. 9-12 - Etienne Barnard, Madelaine Plauché, Marelie H. Davel:
The utility of spoken dialog systems. 13-16 - Roger C. F. Tucker, Mucemi Gakuru:
Experience with developing and deploying an agricultural information system using spoken language technology in Kenya. 17-20 - Neil Patel, Sheetal K. Agarwal, Nitendra Rajput, Amit Anil Nanavati, Paresh Dave, Tapan S. Parikh:
Experiences designing a voice interface for rural India. 21-24
Spoken Language Generation
- Masato Onishi, Toru Takahashi, Toshio Irino, Hideki Kawahara:
Vowel-based frequency alignment function design and recognition-based time alignment for automatic speech morphing. 25-28 - Y. R. Venugopalakrishna, M. V. Vinodh, Hema A. Murthy, Coimbatore S. Ramalingam:
Methods for improving the quality of syllable based speech synthesis. 29-32 - Keikichi Hirose, Qinghua Sun, Nobuaki Minematsu:
Corpus-based synthesis of Mandarin speech with F0 contours generated by superposing tone components on rule-generated phrase components. 33-36 - E. Veera Raghavendra, B. Yegnanarayana, Kishore Prahallad:
Speech synthesis using approximate matching of syllables. 37-40 - Luís Pinto Coelho, Daniela Braga:
Adaptive filtering for high quality hmm based speech synthesis. 41-44 - Marcel Kockmann, Lukás Burget:
Contour modeling of prosodic and acoustic features for speaker recognition. 45-48 - E. Veera Raghavendra, Srinivas Desai, B. Yegnanarayana, Alan W. Black, Kishore Prahallad:
Global syllable set for building speech synthesis in Indian languages. 49-52 - Munkhtuya Davaatsagaan, Kuldip K. Paliwal:
A response generation in the Mongolian spoken language system for accessing to multimedia knowledge base. 53-56
Spoken Language Understanding
- Kinfe Tadesse Mengistu, Mirko Hannemann, Tobias Baum, Andreas Wendemuth:
Hierarchical HMM-based semantic concept labeling model. 57-60 - Marco Dinarelli, Alessandro Moschitti, Giuseppe Riccardi:
Joint generative and discriminative models for spoken language understanding. 61-64 - Arianna Bisazza, Marco Dinarelli, Silvia Quarteroni, Sara Tonelli, Alessandro Moschitti, Giuseppe Riccardi:
Semantic annotations for conversational speech: From speech transcriptions to predicate argument structures. 65-68 - Gökhan Tür, Andreas Stolcke, L. Lynn Voss, John Dowding, Benoît Favre, Raquel Fernández, Matthew Frampton, Michael W. Frandsen, Clint Frederickson, Martin Graciarena, Dilek Hakkani-Tür, Donald Kintzing, Kyle Leveque, Shane Mason, John Niekrasz, Stanley Peters, Matthew Purver, Korbinian Riedhammer, Elizabeth Shriberg, Jing Tien, Dimitra Vergyri, Fan Yang:
The CALO meeting speech recognition and understanding system. 69-72 - Bonaventura Coppola, Alessandro Moschitti, Sara Tonelli, Giuseppe Riccardi:
Automatic framenet-based annotation of conversational speech. 73-76 - Benoît Favre, Dilek Hakkani-Tür, Slav Petrov, Dan Klein:
Efficient sentence segmentation using syntactic features. 77-80 - Kornel Laskowski:
Modeling vocal interaction for text-independent detection of involvement hotspots in multi-party meetings. 81-84 - César González Ferreras, Valentín Cardeñoso-Payo, Emilio Sanchis Arnal:
Experiments in speech driven question answering. 85-88 - John Lee, Stephanie Seneff:
An analysis of grammatical errors in non-native speech in english. 89-92 - Jan Hajic, Silvie Cinková, Marie Mikulová, Petr Pajas, Jan Ptácek, Josef Toman, Zdenka Uresová:
PDTSL: An annotated resource for speech reconstruction. 93-96 - Maria Georgescul, Manny Rayner, Pierrette Bouillon, Nikos Tsourakis:
Discriminative learning using linguistic features to rescore n-best speech hypotheses. 97-100 - Leonardo Badino, Robert A. J. Clark:
Automatic labeling of contrastive word pairs from spontaneous spoken english. 101-104
Spoken Dialog Systems
- David Griol, Lluís F. Hurtado, Encarna Segarra, Emilio Sanchis:
Quantitative evaluation of dialog corpora acquired through different techniques. 105-108 - Ryuichiro Higashinaka, Kohji Dohsaka, Hideki Isozaki:
Effects of self-disclosure and empathy in human-computer dialogue. 109-112 - Hartwig Holzapfel, Alex Waibel:
Modelling multimodal user ID in dialogue. 113-116 - Felix Putze, Hartwig Holzapfel:
IslEnquirer: Social user model acquisition through network analysis and interactive learning. 117-120 - Simon Keizer, Milica Gasic, François Mairesse, Blaise Thomson, Kai Yu, Steve J. Young:
Modelling user behaviour in the HIS-POMDP dialogue manager. 121-124 - David Suendermann, Jackson Liscombe, Keelan Evanini, Krishna Dayanidhi, Roberto Pieraccini:
C5. 125-128 - Keelan Evanini, Phillip Hunter, Jackson Liscombe, David Suendermann, Krishna Dayanidhi, Roberto Pieraccini:
Caller Experience: A method for evaluating dialog systems and its automatic prediction. 129-132 - Dan Bohus, Geoffrey Zweig, Patrick Nguyen, Xiao Li:
Joint n-best rescoring for repeated utterances in spoken dialog systems. 133-136 - Fernando Fernández Martínez, Juan Blázquez, Javier Ferreiros, Roberto Barra-Chicote, Javier Macías Guarasa, Juan Manuel Lucas-Cuesta:
Evaluation of a spoken dialogue system for controlling a Hifi audio system. 137-140 - Géraldine Damnati, Frédéric Béchet, Renato de Mori:
Speaker turn characterization for spoken dialog system monitoring and adaptation. 141-144 - Filipe M. Martins, Joana Paulo Pardal, Luís Franqueira, Pedro Arez, Nuno J. Mamede:
Starting to cook a tutoring dialogue system. 145-148 - Minako Sawaki, Yasuhiro Minami, Ryuichiro Higashinaka, Kohji Dohsaka, Eisaku Maeda:
"Who is this" quiz dialogue system and users' evaluation. 149-152
Speech Document Summarization
- Korbinian Riedhammer, Benoît Favre, Dilek Hakkani-Tür:
A keyphrase based approach to interactive meeting summarization. 153-156 - Shasha Xie, Yang Liu, Hui Lin:
Evaluating the effectiveness of features and sampling in extractive meeting summarization. 157-160 - Justin Jian Zhang, Shilei Huang, Pascale Fung:
RSHMM++ for extractive lecture speech summarization. 161-164 - Sheng-yi Kong, Chien-Chih Wang, Ko-chien Kuo, Lin-Shan Lee:
Automatic title generation for Chinese spoken documents with a delicate scored Viterbi algorithm. 165-168 - Ricardo Ribeiro, David Martins de Matos:
Using prior knowledge to assess relevance in speech summarization. 169-172 - Xiaodan Zhu, Siavash Kazemian, Gerald Penn:
Identifying salient utterances of online spoken documents using descriptive hypertext. 173-176 - Satanjeev Banerjee, Alexander I. Rudnicky:
An extractive-summarization baseline for the automatic detection of noteworthy utterances in multi-party human-human dialog. 177-180 - Fei Liu, Feifan Liu, Yang Liu:
Automatic keyword extraction for the meeting corpus using supervised approach and bigram expansion. 181-184 - Melissa Sherman, Yang Liu:
Using hidden Markov models for topic segmentation of meeting transcripts. 185-188
Topics in Speech and Language Modeling
- Ilya Oparin, Ondrej Glembek, Lukás Burget, Jan Cernocký:
Morphological random forests for language modeling of inflectional languages. 189-192 - Chuang-Hua Chueh, Jen-Tzung Chien:
Continuous topic language modeling for speech recognition. 193-196 - Gilles Boulianne, Maryse Boisvert, Frédéric Osterrath:
Real-time speech recognition captioning of events and meetings. 197-200 - Jen-Tzung Chien, Chuang-Hua Chueh:
Latent dirichlet language model for speech recognition. 201-204 - Hany Hassan, Khalil Sima'an, Andy Way:
A syntactic language model based on incremental CCG parsing. 205-208 - Ryuichi Nisimura, Jumpei Miyake, Hideki Kawahara, Toshio Irino:
Speech-to-text input method for web system using JavaScript. 209-212 - Richard Dufour, Yannick Estève:
Correcting asr outputs: Specific solutions to specific errors in French. 213-216 - Vincent Wan, John Dines, Asmaa El Hannani, Thomas Hain:
Bob: A lexicon and pronunciation dictionary generator. 217-220 - Shilei Huang, Xiang Xie, Pascale Fung:
Using output probability distribution for oov word rejection. 221-224 - Kaustubh Kulkarni, Sohini Sengupta, V. Ramasubramanian, Josef G. Bauer, Georg Stemmer:
Accented Indian english ASR: Some early results. 225-228 - Konakanchi Partha Sarathy, A. G. Ramakrishnan:
A research bed for unit selection based text to speech synthesis. 229-232
Spoken Language Translation
- Matthias Wölfel, Muntsin Kolss, Florian Kraft, Jan Niehues, Matthias Paulik, Alex Waibel:
Simultaneous machine translation of german lectures into english: Investigating research challenges for the future. 233-236 - Khalil Sima'an, Markos Mylonakis:
Better statistical estimation can benefit all phrases in phrase-based statistical machine translation. 237-240 - Paisarn Charoenpornsawat, Tanja Schultz:
Improving word segmentation for Thai speech translation. 241-244 - Fred Choi, Stavros Tsakalidis, Shirin Saleem, Chia-Lin Kao, Ralf Meermeier, Kriste Krstovski, Christine Moran, Krishna Subramanian, Rohit Prasad, Prem Natarajan:
Recent improvements in BBN's English/Iraqi speech-to-speech translation system. 245-248 - Rohit Prasad, Christine Moran, Fred Choi, Ralf Meermeier, Shirin Saleem, Chia-Lin Kao, David Stallard, Prem Natarajan:
Name aware speech-to-speech translation for English/Iraqi. 249-252 - Sameer Maskey, Martin Cmejrek, Bowen Zhou, Yuqing Gao:
Class-based named entity translation in a speech to speech translation system. 253-256 - Damianos G. Karakos, Sanjeev Khudanpur:
Sequential system combination for machine translation of speech. 257-260 - Arindam Mandal, Dimitra Vergyri, Wen Wang, Jing Zheng, Andreas Stolcke, Gökhan Tür, Dilek Hakkani-Tür, Necip Fazil Ayan:
Efficient data selection for machine translation. 261-264 - A. Ryan Aminzadeh, Wade Shen:
Low-resource speech translation of Urdu to English using semi-supervised part-of-speech tagging and transliteration. 265-268 - Vivek Kumar Rangarajan Sridhar, Shrikanth S. Narayanan, Srinivas Bangalore:
Incorporating discourse context in spoken language translation through dialog acts. 269-272
Spoken Document Retrieval
- Igor Szöke, Lukás Burget, Jan Cernocký, Michal Fapso:
Sub-word modeling of out of vocabulary words in spoken term detection. 273-276 - Fernando Batista, Rui Amaral, Isabel Trancoso, Nuno J. Mamede:
Impact of dynamic model adaptation beyond speech recognition. 277-280 - Heng Ji, Ralph Grishman, Wen Wang:
Phonetic name matching for cross-lingual Spoken Sentence Retrieval. 281-284 - Hung-lin Chang, Yi-Cheng Pan, Lin-Shan Lee:
Latent semantic retrieval of spoken documents over position specific posterior lattices. 285-288 - Yi-Cheng Pan, Hung-lin Chang, Lin-Shan Lee:
Robustness analysis on lattice-based speech indexing approaches with respect to varying recognition accuracies by refined simulations. 289-292 - Frank Seide, Kishan Thambiratnam, Roger Peng Yu:
Word-lattice based spoken-document indexing with standard text indexers. 293-296 - Junta Mizuno, Jun Ogata, Masataka Goto:
A similar content retrieval method for podcast episodes. 297-300 - Go Kuriki, Yoshiaki Itoh, Kazunori Kojima, Masaaki Ishigame, Kazuyo Tanaka, Shi-wook Lee:
Open vocabulary spoken document retrieval by subword sequence obtained from speech recognizer. 301-304 - Mickael Rouvier, Georges Linarès, Benjamin Lecouteux:
On-the-fly term spotting by phonetic filtering and request-driven decoding. 305-308 - Kamini Malhotra, Anu Khosla:
Automatic identification of gender & accent in spoken Hindi utterances with regional Indian accents. 309-312 - Manish Gaurav:
Performance analysis of spectral and prosodic features and their fusion for emotion recognition in speech. 313-316
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.