default search action
Keiichiro Oura
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2023
- [c56]Takenori Yoshimura, Shinji Takaki, Kazuhiro Nakamura, Keiichiro Oura, Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda:
Embedding a Differentiable Mel-Cepstral Synthesis Filter to a Neural Speech Synthesis System. ICASSP 2023: 1-5 - [c55]Takenori Yoshimura, Takato Fujimoto, Keiichiro Oura, Keiichi Tokuda:
SPTK4: An Open-Source Software Toolkit for Speech Signal Processing. SSW 2023: 211-217 - 2022
- [i6]Takenori Yoshimura, Shinji Takaki, Kazuhiro Nakamura, Keiichiro Oura, Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda:
Embedding a Differentiable Mel-cepstral Synthesis Filter to a Neural Speech Synthesis System. CoRR abs/2211.11222 (2022) - 2021
- [j11]Yukiya Hono, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
PeriodNet: A Non-Autoregressive Raw Waveform Generative Model With a Structure Separating Periodic and Aperiodic Components. IEEE Access 9: 137599-137612 (2021) - [j10]Yukiya Hono, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System. IEEE ACM Trans. Audio Speech Lang. Process. 29: 2803-2815 (2021) - [c54]Yukiya Hono, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Periodnet: A Non-Autoregressive Waveform Generation Model with a Structure Separating Periodic and Aperiodic Components. ICASSP 2021: 6049-6053 - [i5]Yukiya Hono, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
PeriodNet: A non-autoregressive waveform generation model with a structure separating periodic and aperiodic components. CoRR abs/2102.07786 (2021) - [i4]Yukiya Hono, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System. CoRR abs/2108.02776 (2021) - 2020
- [c53]Kazuhiro Nakamura, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Fast and High-Quality Singing Voice Synthesis System Based on Convolutional Neural Networks. ICASSP 2020: 7239-7243 - [c52]Takato Fujimoto, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Semi-Supervised Learning Based on Hierarchical Generative Models for End-to-End Speech Synthesis. ICASSP 2020: 7644-7648 - [c51]Yukiya Hono, Kazuna Tsuboi, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Hierarchical Multi-Grained Generative Model for Expressive Speech Synthesis. INTERSPEECH 2020: 3441-3445 - [i3]Yukiya Hono, Kazuna Tsuboi, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Hierarchical Multi-Grained Generative Model for Expressive Speech Synthesis. CoRR abs/2009.08474 (2020)
2010 – 2019
- 2019
- [c50]Yukiya Hono, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Singing Voice Synthesis Based on Generative Adversarial Networks. ICASSP 2019: 6955-6959 - [c49]Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Speaker-dependent Wavenet-based Delay-free Adpcm Speech Coding. ICASSP 2019: 7145-7149 - [c48]Keiichiro Oura, Kazuhiro Nakamura, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda:
Deep neural network based real-time speech vocoder with periodic and aperiodic inputs. SSW 2019: 13-18 - [c47]Takato Fujimoto, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Impacts of input linguistic feature representation on Japanese end-to-end speech synthesis. SSW 2019: 166-171 - [c46]Motoki Shimada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Low computational cost speech synthesis based on deep neural networks using hidden semi-Markov model structures. SSW 2019: 177-182 - [i2]Kazuhiro Nakamura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Singing voice synthesis based on convolutional neural networks. CoRR abs/1904.06868 (2019) - [i1]Kazuhiro Nakamura, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Fast and High-Quality Singing Voice Synthesis System based on Convolutional Neural Networks. CoRR abs/1910.11690 (2019) - 2018
- [j9]Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Mel-Cepstrum-Based Quantization Noise Shaping Applied to Neural-Network-Based Speech Waveform Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 26(7): 1173-1180 (2018) - [c45]Takayuki Kasugai, Yoshinari Tsuzuki, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Image Recognition Based on Convolutional Neural Networks Using Features Generated from Separable Lattice Hidden Markov Models. APSIPA 2018: 324-328 - [c44]Kento Nakao, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Speaker Adaptation for Speech Synthesis Based on Deep Neural Networks Using Hidden Semi-Markov Model Structures. APSIPA 2018: 638-643 - [c43]Takato Fujimoto, Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Speech Synthesis Using WaveNet Vocoder Based on Periodic/Aperiodic Decomposition. APSIPA 2018: 644-648 - [c42]Yukiya Hono, Shumma Murata, Kazuhiro Nakamura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Recent Development of the DNN-based Singing Voice Synthesis System - Sinsy. APSIPA 2018: 1003-1009 - [c41]Takenori Yoshimura, Natsumi Koike, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Discriminative Feature Extraction Based on Sequential Variational Autoencoder for Speaker Recognition. APSIPA 2018: 1742-1746 - [c40]Koki Senda, Yukiya Hono, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Singing Voice Conversion Using Posted Waveform Data on Music Social Media. APSIPA 2018: 1913-1917 - [c39]Kei Sawada, Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
The NITech text-to-speech system for the Blizzard Challenge 2018. Blizzard Challenge 2018 - [c38]Jumpei Niwa, Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Statistical Voice Conversion Based on Wavenet. ICASSP 2018: 5289-5293 - [c37]Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
WaveNet-Based Zero-Delay Lossless Speech Coding. SLT 2018: 153-158 - 2017
- [j8]Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Simultaneous Optimization of Multiple Tree-Based Factor Analyzed HMM for Speech Synthesis. IEEE ACM Trans. Audio Speech Lang. Process. 25(9): 1836-1845 (2017) - [c36]Anocha Rugchatjaroen, Sittipong Saychum, Keiichiro Oura, Keiichi Tokuda:
Generalization of Thai tone contour in HMM-based speech synthesis. APSIPA 2017: 1102-1105 - [c35]Kei Sawada, Kei Hashimoto, Keiichiro Oura, Keiichi Tokuda:
The NITech text-to-speech system for the Blizzard Challenge 2017. Blizzard Challenge 2017 - [p1]Keiichi Tokuda, Akinobu Lee, Yoshihiko Nankaku, Keiichiro Oura, Kei Hashimoto, Daisuke Yamamoto, Ichi Takumi, Takahiro Uchiya, Shuhei Tsutsumi, Steve Renals, Junichi Yamagishi:
User Generated Dialogue Systems: uDialogue. Human-Harmonized Information Technology (2) 2017: 77-114 - 2016
- [c34]Kei Sawada, Chiaki Asai, Kei Hashimoto, Keiichiro Oura, Keiichi Tokuda:
The NITech text-to-speech system for the Blizzard Challenge 2016. Blizzard Challenge 2016 - [c33]Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Trajectory training considering global variance for speech synthesis based on neural networks. ICASSP 2016: 5600-5604 - [c32]Naoki Hosaka, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Voice Conversion Based on Trajectory Model Training of Neural Networks Considering Global Variance. INTERSPEECH 2016: 307-311 - [c31]Masanari Nishimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Singing Voice Synthesis Based on Deep Neural Networks. INTERSPEECH 2016: 2478-2482 - [c30]Rasmus Dall, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Redefining the Linguistic Context Feature Set for HMM and DNN TTS Through Position and Parsing. INTERSPEECH 2016: 2851-2855 - [c29]Keiichi Tokuda, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku:
Temporal modeling in neural network based statistical parametric speech synthesis. SSW 2016: 106-111 - 2015
- [j7]Hongwu Yang, Keiichiro Oura, Haiyan Wang, Zhenye Gan, Keiichi Tokuda:
Using speaker adaptive training to realize Mandarin-Tibetan cross-lingual speech synthesis. Multim. Tools Appl. 74(22): 9927-9942 (2015) - [c28]Kei Sawada, Kei Hashimoto, Keiichiro Oura, Keiichi Tokuda:
The NITECH HMM-based text-to-speech system for the Blizzard Challenge 2015. Blizzard Challenge 2015 - [c27]Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
The effect of neural networks in statistical parametric speech synthesis. ICASSP 2015: 4455-4459 - 2014
- [c26]Kei Sawada, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Keiichi Tokuda:
Overview of NITECH HMM-based text-to-speech system for Blizzard Challenge 2014. Blizzard Challenge 2014 - [c25]Daisuke Yamamoto, Keiichiro Oura, Ryota Nishimura, Takahiro Uchiya, Akinobu Lee, Ichi Takumi, Keiichi Tokuda:
Voice interaction system with 3D-CG virtual agent for stand-alone smartphones. HAI 2014: 323-330 - [c24]Kazuhiro Nakamura, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
HMM-Based singing voice synthesis and its application to Japanese and English. ICASSP 2014: 265-269 - [c23]Kanako Shirota, Kazuhiro Nakamura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Integration of speaker and pitch adaptive training for HMM-based singing voice synthesis. ICASSP 2014: 2559-2563 - [c22]Kazuhiro Nakamura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
A mel-cepstral analysis technique restoring high frequency components from low-sampling-rate speech. INTERSPEECH 2014: 2494-2498 - 2013
- [j6]John Dines, Hui Liang, Lakshmi Saheer, Matthew Gibson, William Byrne, Keiichiro Oura, Keiichi Tokuda, Junichi Yamagishi, Simon King, Mirjam Wester, Teemu Hirsimäki, Reima Karhila, Mikko Kurimo:
Personalising speech-to-speech translation: Unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis. Comput. Speech Lang. 27(2): 420-437 (2013) - [j5]Keiichi Tokuda, Yoshihiko Nankaku, Tomoki Toda, Heiga Zen, Junichi Yamagishi, Keiichiro Oura:
Speech Synthesis Based on Hidden Markov Models. Proc. IEEE 101(5): 1234-1252 (2013) - [c21]Hongwu Yang, Keiichiro Oura, Zhenye Gan, Keiichi Tokuda:
Realizing Tibetan speech synthesis by speaker adaptive training. APSIPA 2013: 1-4 - [c20]Shinji Takaki, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Keiichi Tokuda:
Overview of NITECH HMM-based speech synthesis system for Blizzard Challenge 2013. Blizzard Challenge 2013 - [c19]Akinobu Lee, Keiichiro Oura, Keiichi Tokuda:
Mmdagent - A fully open-source toolkit for voice interaction systems. ICASSP 2013: 8382-8385 - [c18]Christophe Veaux, Maria Astrinaki, Keiichiro Oura, Robert A. J. Clark, Junichi Yamagishi:
Real-time control of expressive speech synthesis using kinect body tracking. SSW 2013: 247-248 - [c17]Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
Cross-lingual speaker adaptation based on factor analysis using bilingual speech data for HMM-based speech synthesis. SSW 2013: 297-302 - 2012
- [j4]Keiichiro Oura, Junichi Yamagishi, Mirjam Wester, Simon King, Keiichi Tokuda:
Analysis of unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis using KLD-based transform mapping. Speech Commun. 54(6): 703-714 (2012) - [c16]Shinji Takaki, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Keiichi Tokuda:
Overview of NIT HMM-based speech synthesis system for Blizzard Challenge 2012. Blizzard Challenge 2012 - [c15]Keiichiro Oura, Ayami Mase, Yoshihiko Nankaku, Keiichi Tokuda:
Pitch adaptive training for hmm-based singing voice synthesis. ICASSP 2012: 5377-5380 - 2011
- [c14]Kei Hashimoto, Shinji Takaki, Keiichiro Oura, Keiichi Tokuda:
Overview of NIT HMM-based speech synthesis system for Blizzard Challenge 2011. Blizzard Challenge 2011 - [c13]Shinji Takaki, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
An optimization algorithm of independent mean and variance parameter tying structures for HMM-based speech synthesis. ICASSP 2011: 4700-4703 - [c12]Tsuneo Kato, Makoto Yamada, Nobuyuki Nishizawa, Keiichiro Oura, Keiichi Tokuda:
Large-Scale Subjective Evaluations of Speech Rate Control Methods for HMM-Based Speech Synthesizers. INTERSPEECH 2011: 1845-1848 - 2010
- [j3]Keiichiro Oura, Heiga Zen, Yoshihiko Nankaku, Akinobu Lee, Keiichi Tokuda:
A Covariance-Tying Technique for HMM-Based Speech Synthesis. IEICE Trans. Inf. Syst. 93-D(3): 595-601 (2010) - [j2]Junichi Yamagishi, Bela Usabaev, Simon King, Oliver Watts, John Dines, Jilei Tian, Yong Guan, Rile Hu, Keiichiro Oura, Yi-Jian Wu, Keiichi Tokuda, Reima Karhila, Mikko Kurimo:
Thousands of Voices for HMM-Based Speech Synthesis-Analysis and Application of TTS Systems Built on Various ASR Corpora. IEEE Trans. Speech Audio Process. 18(5): 984-1004 (2010) - [c11]Mikko Kurimo, William Byrne, John Dines, Philip N. Garner, Matthew Gibson, Yong Guan, Teemu Hirsimäki, Reima Karhila, Simon King, Hui Liang, Keiichiro Oura, Lakshmi Saheer, Matt Shannon, Sayaka Shiota, Jilei Tian:
Personalising Speech-To-Speech Translation in the EMIME Project. ACL (System Demonstrations) 2010: 48-53 - [c10]Keiichiro Oura, Kei Hashimoto, Sayaka Shiota, Keiichi Tokuda:
Overview of NIT HMM-based speech synthesis system for Blizzard Challenge 2010. Blizzard Challenge 2010 - [c9]Keiichiro Oura, Keiichi Tokuda, Junichi Yamagishi, Simon King, Mirjam Wester:
Unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis. ICASSP 2010: 4594-4597 - [c8]Ayami Mase, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda:
HMM-based singing voice synthesis system using pitch-shifted pseudo training data. INTERSPEECH 2010: 845-848 - [c7]Mirjam Wester, John Dines, Matthew Gibson, Hui Liang, Yi-Jian Wu, Lakshmi Saheer, Simon King, Keiichiro Oura, Philip N. Garner, William Byrne, Yong Guan, Teemu Hirsimäki, Reima Karhila, Mikko Kurimo, Matt Shannon, Sayaka Shiota, Jilei Tian, Keiichi Tokuda, Junichi Yamagishi:
Speaker adaptation and the evaluation of speaker similarity in the EMIME speech-to-speech translation project. SSW 2010: 192-197 - [c6]Keiichiro Oura, Ayami Mase, Tomohiko Yamada, Satoru Muto, Yoshihiko Nankaku, Keiichi Tokuda:
Recent development of the HMM-based singing voice synthesis system - Sinsy. SSW 2010: 211-216
2000 – 2009
- 2009
- [c5]Keiichiro Oura, Yi-Jian Wu, Keiichi Tokuda:
Overview of NIT HMM-based speech synthesis system for Blizzard Challenge 2009. Blizzard Challenge 2009 - [c4]Junichi Yamagishi, Bela Usabaev, Simon King, Oliver Watts, John Dines, Jilei Tian, Rile Hu, Yong Guan, Keiichiro Oura, Keiichi Tokuda, Reima Karhila, Mikko Kurimo:
Thousands of voices for HMM-based speech synthesis. INTERSPEECH 2009: 420-423 - [c3]Keiichiro Oura, Heiga Zen, Yoshihiko Nankaku, Akinobu Lee, Keiichi Tokuda:
Tying covariance matrices to reduce the footprint of HMM-based speech synthesis systems. INTERSPEECH 2009: 1759-1762 - 2008
- [j1]Keiichiro Oura, Heiga Zen, Yoshihiko Nankaku, Akinobu Lee, Keiichi Tokuda:
A Fully Consistent Hidden Semi-Markov Model-Based Speech Recognition System. IEICE Trans. Inf. Syst. 91-D(11): 2693-2700 (2008) - [c2]Keiichiro Oura, Yoshihiko Nankaku, Tomoki Toda, Keiichi Tokuda, Ranniery Maia, Shinsuke Sakai, Satoshi Nakamura:
Simultaneous Acoustic, Prosodic, and Phrasing Model Training for TTs Conversion Systems. ISCSLP 2008: 1-4 - 2006
- [c1]Keiichiro Oura, Heiga Zen, Yoshihiko Nankaku, Akinobu Lee, Keiichi Tokuda:
Hidden Semi-Markov Model Based Speech Recognition System using Weighted Finite-State Transducer. ICASSP (1) 2006: 33-36
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-10-07 21:20 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint