default search action
Ben P. Milner
Person information
- affiliation: School of Computer Science, University of East Anglia, Norwich, UK
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2023
- [c85]Georgiana-Elena Sfeclis, Ben Milner, Danny Websdale:
Investigating Imaginary Mask Estimation in Complex Masking for Speech Enhancement. EUSIPCO 2023: 131-135 - [c84]Jonathan Windle, Iain A. Matthews, Ben Milner, Sarah Taylor:
The UEA Digital Humans entry to the GENEA Challenge 2023. ICMI 2023: 802-810 - 2022
- [j19]Danny Websdale, Sarah Taylor, Ben Milner:
Speaker-Independent Speech Animation Using Perceptual Loss Functions and Synthetic Data. IEEE Trans. Multim. 24: 2539-2552 (2022) - 2021
- [c83]William Vickers, Ben Milner, Robert Lee:
Improving The Robustness Of Right Whale Detection In Noisy Conditions Using Denoising Autoencoders And Augmented Training. ICASSP 2021: 91-95
2010 – 2019
- 2019
- [j18]Ausdang Thangthai, Ben Milner, Sarah Taylor:
Synthesising visual speech using dynamic visemes and deep learning architectures. Comput. Speech Lang. 55: 101-119 (2019) - 2018
- [j17]Faheem Khan, Ben P. Milner, Thomas Le Cornu:
Using Visual Speech Information in Masking Methods for Audio Speaker Separation. IEEE ACM Trans. Audio Speech Lang. Process. 26(10): 1742-1754 (2018) - [c82]Danny Websdale, Sarah Taylor, Ben Milner:
The Effect of Real-Time Constraints on Automatic Speech Animation. INTERSPEECH 2018: 2479-2483 - 2017
- [j16]Philip Harding, Ben Milner:
Estimating acoustic speech features in low signal-to-noise ratios using a statistical framework. Comput. Speech Lang. 42: 1-19 (2017) - [j15]Thomas Le Cornu, Ben Milner:
Generating Intelligible Audio Speech From Visual Speech. IEEE ACM Trans. Audio Speech Lang. Process. 25(9): 1751-1761 (2017) - [c81]Danny Websdale, Ben Milner:
Using visual speech information and perceptually motivated loss functions for binary mask estimation. AVSP 2017: 41-46 - [c80]Danny Websdale, Ben Milner:
A Comparison of Perceptually Motivated Loss Functions for Binary Mask Estimation in Speech Separation. INTERSPEECH 2017: 2003-2007 - 2016
- [c79]Sarah Taylor, Akihiro Kato, Iain A. Matthews, Ben P. Milner:
Audio-to-Visual Speech Conversion Using Deep Neural Networks. INTERSPEECH 2016: 1482-1486 - [c78]Ausdang Thangthai, Ben Milner, Sarah Taylor:
Visual Speech Synthesis Using Dynamic Visemes, Contextual Features and DNNs. INTERSPEECH 2016: 2458-2462 - [c77]Akihiro Kato, Ben P. Milner:
HMM-Based Speech Enhancement Using Sub-Word Models and Noise Adaptation. INTERSPEECH 2016: 3748-3752 - 2015
- [j14]Philip Harding, Ben Milner:
Reconstruction-based speech enhancement from robust acoustic features. Speech Commun. 75: 62-75 (2015) - [c76]Thomas Le Cornu, Ben Milner:
Voicing classification of visual speech using convolutional neural networks. AVSP 2015: 103-108 - [c75]Danny Websdale, Ben Milner:
Analysing the importance of different visual feature coefficients. AVSP 2015: 137-142 - [c74]Danny Websdale, Thomas Le Cornu, Ben Milner:
Objective measures for predicting the intelligibility of spectrally smoothed speech with artificial excitation. INTERSPEECH 2015: 638-642 - [c73]Faheem Khan, Ben Milner:
Using audio and visual information for single channel speaker separation. INTERSPEECH 2015: 1517-1521 - [c72]Thomas Le Cornu, Ben Milner:
Reconstructing intelligible audio speech from visual speech features. INTERSPEECH 2015: 3355-3359 - 2014
- [c71]Akihiro Kato, Ben Milner:
Using hidden Markov models for speech enhancement. INTERSPEECH 2014: 2695-2699 - 2013
- [j13]Ben Milner:
Enhancing speech at very low signal-to-noise ratios using non-acoustic reference signals. Speech Commun. 55(9): 879-892 (2013) - [c70]Faheem Khan, Ben Milner:
Speaker separation using visually-derived binary masks. AVSP 2013: 215-220 - [c69]John H. Taylor, Ben Milner:
Modelling and estimation of the fundamental frequency of speech using a hidden Markov model. INTERSPEECH 2013: 1926-1930 - [c68]Faheem Khan, Ben Milner:
Speaker separation using visual speech features and single-channel audio. INTERSPEECH 2013: 3264-3268 - 2012
- [c67]Philip Harding, Ben Milner:
Enhancing Speech by Reconstruction from Robust Acoustic Features. INTERSPEECH 2012: 943-946 - [c66]Philip Harding, Ben Milner:
On the use of Machine Learning Methods for Speech and Voicing Classification. INTERSPEECH 2012: 1981-1984 - 2011
- [j12]Ben Milner, Jonathan Darch:
Robust Acoustic Speech Feature Prediction From Noisy Mel-Frequency Cepstral Coefficients. IEEE Trans. Speech Audio Process. 19(2): 338-347 (2011) - [j11]Ibrahim Almajai, Ben Milner:
Visually Derived Wiener Filters for Speech Enhancement. IEEE Trans. Speech Audio Process. 19(6): 1642-1651 (2011) - [c65]Ben Milner:
Maximum a posteriori Estimation of Noise from Non-Acoustic Reference Signals in Very Low Signal-to-Noise Ratio Environments. INTERSPEECH 2011: 357-360 - [c64]Philip Harding, Ben Milner:
Speech Enhancement by Reconstruction from Cleaned Acoustic Features. INTERSPEECH 2011: 1189-1192 - [c63]Alipah Pawi, Saeed Vaseghi, Ben Milner, Seyed Ghorshi:
Fundamental Frequency Estimation Using Modified Higher Order Moments and Multiple Windows. INTERSPEECH 2011: 1965-1968 - 2010
- [c62]Alipah Pawi, Saeed Vaseghi, Ben Milner:
Pitch extraction using modified higher order moments. ICASSP 2010: 5078-5081
2000 – 2009
- 2009
- [c61]Ibrahim Almajai, Ben Milner:
Effective visually-derived Wiener filtering for audio-visual speech processing. AVSP 2009: 134-139 - [c60]Ben Milner, Jonathan Darch, Ibrahim Almajai:
Reconstructing clean speech from noisy MFCC vectors. INTERSPEECH 2009: 1943-1946 - [c59]Ibrahim Almajai, Ben Milner:
Enhancing audio speech using visual speech features. INTERSPEECH 2009: 1959-1962 - 2008
- [j10]Qin Yan, Saeed Vaseghi, Esfandiar Zavarehei, Ben Milner, Jonathan Darch, Paul R. White, Ioannis Andrianakis:
Kalman tracking of linear predictor and harmonic noise models for noisy speech enhancement. Comput. Speech Lang. 22(1): 69-83 (2008) - [c58]Ibrahim Almajai, Ben P. Milner:
Using audio-visual features for robust voice activity detection in clean and noisy speech. EUSIPCO 2008: 1-5 - [c57]Ben P. Milner, Jonathan Darch, Ibrahim Almajai, Saeed Vaseghi:
Comparing noise compensation methods for robust prediction of acoustic speech features from MFCC vectors in noise. EUSIPCO 2008: 1-5 - [c56]Ben Milner, Jonathan Darch, Saeed Vaseghi:
Applying noise compensation methods to robustly predict acoustic speech features from MFCC vectors in noise. ICASSP 2008: 3945-3948 - 2007
- [j9]Qin Yan, Saeed Vaseghi, Esfandiar Zavarehei, Ben P. Milner, Jonathan Darch, Paul R. White, Ioannis Andrianakis:
Formant tracking linear prediction model using HMMs and Kalman filters for noisy speech processing. Comput. Speech Lang. 21(3): 543-561 (2007) - [j8]Ben Milner, Xu Shao:
Prediction of Fundamental Frequency and Voicing From Mel-Frequency Cepstral Coefficients for Unconstrained Speech Reconstruction. IEEE Trans. Speech Audio Process. 15(1): 24-33 (2007) - [c55]Ben Milner, Ibrahim Almajai:
Noisy audio speech enhancement using Wiener filters derived from visual speech. AVSP 2007: 16 - [c54]Ibrahim Almajai, Ben Milner:
Maximising audio-visual speech correlation. AVSP 2007: 17 - [c53]Qin Yan, Saeed Vaseghi, Esfandiar Zavarehei, Ben P. Milner:
Restoration of noisy and band limited archived speech records with linear predictor and harmonic noise models. EUSIPCO 2007: 80-84 - [c52]Jonathan Darch, Ben Milner, Ibrahim Almajai, Saeed Vaseghi:
An Investigation into the Correlation and Prediction of Acoustic Speech Features from MFCC Vectors. ICASSP (4) 2007: 465-468 - [c51]Ibrahim Almajai, Ben Milner, Jonathan Darch, Saeed Vaseghi:
Visually-Derived Wiener Filters for Speech Enhancement. ICASSP (4) 2007: 585-588 - [c50]Jonathan Darch, Ben Milner:
A comparison of estimated and MAP-predicted formants and fundamental frequencies with a speech reconstruction application. INTERSPEECH 2007: 542-545 - 2006
- [j7]Ben Milner, Xu Shao:
Clean speech reconstruction from MFCC vectors and fundamental frequency using an integrated front-end. Speech Commun. 48(6): 697-715 (2006) - [j6]Ben Milner, Christian Wellekens, Børge Lindberg:
Special Issue on Robustness Issues for Conversational Interaction. Speech Commun. 48(11): 1399-1401 (2006) - [j5]Alastair Bruce James, Ben Milner:
Towards improving the robustness of distributed speech recognition in packet loss. Speech Commun. 48(11): 1402-1421 (2006) - [j4]Jonathan Darch, Ben P. Milner, Saeed Vaseghi:
MAP prediction of formant frequencies and voicing class from MFCC vectors in noise. Speech Commun. 48(11): 1556-1572 (2006) - [j3]Ben Milner, Alastair Bruce James:
Robust speech recognition over mobile and IP networks in burst-like packet loss. IEEE Trans. Speech Audio Process. 14(1): 223-231 (2006) - [j2]Ling Ma, Ben Milner, Dan J. Smith:
Acoustic environment classification. ACM Trans. Speech Lang. Process. 3(2): 1-22 (2006) - [c49]Qin Yan, Saeed Vaseghi, Esfandiar Zavarehei, Ben P. Milner:
Kalman filter with linear predictor and harmonic noise models for noisy speech enhancement. EUSIPCO 2006: 1-4 - [c48]Mark Hadley, Ben Milner, Richard W. Harvey:
Noise Reduction for Driver-To-Pit-Crew Communication in Motor Racing. ICASSP (1) 2006: 165-168 - [c47]Ibrahim Almajai, Ben Milner, Jonathan Darch:
Analysis of correlation between audio and visual speech features for clean audio feature prediction in noise. INTERSPEECH 2006 - [c46]Jonathan Darch, Ben Milner:
HMM-based MAP prediction of voiced and unvoiced formant frequencies from noisy MFCC vectors. INTERSPEECH 2006 - [c45]Ross Tailby, Richard Dean, Ben Milner, Dan J. Smith:
Email classification for automated service handling. SAC 2006: 1073-1077 - 2005
- [c44]Alastair Bruce James, Ben Milner:
Soft Decoding of Temporal Derivatives for Robust Distributed Speech Recognition in Packet Loss. ICASSP (1) 2005: 345-348 - [c43]Jonathan Darch, Ben Milner, Xu Shao, Saeed Vaseghi, Qin Yan:
Predicting Formant Frequencies from MFCC Vectors. ICASSP (1) 2005: 941-944 - [c42]Ben Milner, Xu Shao, Jonathan Darch:
Fundamental frequency and voicing prediction from MFCCs for speech reconstruction from unconstrained speech. INTERSPEECH 2005: 321-324 - [c41]Jonathan Darch, Ben P. Milner, Saeed Vaseghi:
Formant frequency prediction from MFCC vectors in noisy environments. INTERSPEECH 2005: 1129-1132 - [c40]Qin Yan, Saeed Vaseghi, Esfandiar Zavarehei, Ben P. Milner:
Formant-tracking linear prediction models for speech processing in noisy environments. INTERSPEECH 2005: 2081-2084 - [c39]Alastair Bruce James, Ben Milner:
Combining packet loss compensation methods for robust distributed speech recognition. INTERSPEECH 2005: 2857-2860 - 2004
- [c38]A. B. James, Ben P. Milner:
Interleaving and estimation of lost vectors for robust speech recognition in burst-like packet loss. EUSIPCO 2004: 1947-1950 - [c37]Xu Shao, Ben Milner:
Pitch prediction from MFCC vectors for speech reconstruction. ICASSP (1) 2004: 97-100 - [c36]Alastair Bruce James, Ben P. Milner:
An analysis of interleavers for robust speech recognition in burst-like packet loss. ICASSP (1) 2004: 853-856 - [c35]Alastair Bruce James, Ben P. Milner, Angel Manuel Gomez:
A comparison of packet loss compensation methods and interleaving for speech recognition in burst-like packet loss. INTERSPEECH 2004: 1545-1548 - [c34]Ben P. Milner, Alastair Bruce James:
An analysis of packet loss models for distributed speech recognition. INTERSPEECH 2004: 1549-1552 - [c33]Xu Shao, Ben P. Milner:
MAP prediction of pitch from MFCC vectors for speech reconstruction. INTERSPEECH 2004: 2425-2428 - 2003
- [c32]Ling Ma, Dan J. Smith, Ben Milner:
Environmental Noise Classification for Context-Aware Applications. DEXA 2003: 360-370 - [c31]Ben Milner, Xu Shao:
Low bit-rate feature vector compression using transform coding and non-uniform bit allocation. ICASSP (2) 2003: 129-132 - [c30]Xu Shao, Ben Milner:
Clean speech reconstruction from noisy mel-frequency cepstral coefficients using a sinusoidal model. ICASSP (1) 2003: 704-707 - [c29]Tanya Lambert, Andrew P. Breen, Barry Eggleton, Stephen J. Cox, Ben P. Milner:
Unit selection in concatenative TTS synthesis systems based on mel filter bank amplitudes and phonetic context. INTERSPEECH 2003: 273-276 - [c28]Xu Shao, Ben P. Milner, Stephen J. Cox:
Integrated pitch and MFCC extraction for speech reconstruction and speech recognition applications. INTERSPEECH 2003: 1725-1728 - [c27]Ling Ma, Dan J. Smith, Ben P. Milner:
Context awareness using environmental noise classification. INTERSPEECH 2003: 2237-2240 - [c26]Ben P. Milner, Alastair Bruce James:
Analysis and compensation of packet loss in distributed speech recognition using interleaving. INTERSPEECH 2003: 2693-2696 - [c25]Ben P. Milner:
Non-linear compression of feature vectors using transform coding and non-uniform bit allocation. INTERSPEECH 2003: 2697-2700 - 2002
- [c24]Ben Milner:
A comparison of front-end configurations for robust speech recognition. ICASSP 2002: 797-800 - [c23]Djemal H. Kolonic, Danilo P. Mandic, Ben Milner, Richard W. Harvey:
On the derivation of the optimal payload size for packet based transmission over a binary symmetrical communication channel. ICASSP 2002: 4120-4123 - [c22]Ben Milner, Xu Shao:
Transform-based feature vector compression for distributed speech recognition. INTERSPEECH 2002: 2233-2236 - [c21]Ben Milner, Xu Shao:
Speech reconstruction from mel-frequency cepstral coefficients using a source-filter model. INTERSPEECH 2002: 2421-2424 - 2001
- [c20]Ben Milner:
Robust speech recognition in burst-like packet loss. ICASSP 2001: 261-264 - 2000
- [c19]Ben Milner, Shahram Semnani:
Robust speech recognition over IP networks. ICASSP 2000: 1791-1794 - [c18]Ben Milner:
Robust voice recognition over IP and mobile networks. PIMRC 2000: 1197-1201
1990 – 1999
- 1999
- [c17]Ben Milner, Mark Farrell:
A comparison of techniques for tone compensation in payphone-based speech recognition. EUROSPEECH 1999: 2845-2848 - 1998
- [c16]Daniel Azzopardi, Shahram Semnani, Ben Milner, Richard Wiseman:
Improving accuracy of telephony-based, speaker-independent speech recognition. ICSLP 1998 - [c15]Naomi Harte, Saeed Vaseghi, Ben P. Milner:
Joint recognition and segmentation using phonetically derived features and a hybrid phoneme model. ICSLP 1998 - 1997
- [j1]Saeed Vaseghi, Ben P. Milner:
Noise compensation methods for hidden Markov model speech recognition in adverse environments. IEEE Trans. Speech Audio Process. 5(1): 11-21 (1997) - [c14]Saeed Vaseghi, Naomi Harte, Ben Milner:
Multi-resolution phonetic/segmental features and models for HMM-based speech recognition. ICASSP 1997: 1263-1266 - [c13]Ben P. Milner:
Cepstral-time matrices and LDA for improved connected digit and sub-word recognition accuracy. EUROSPEECH 1997: 405-408 - [c12]Simon Nicholson, Ben P. Milner, Stephen J. Cox:
Evaluating feature set performance using the f-ratio and j-measures. EUROSPEECH 1997: 413-416 - 1996
- [c11]Ben Milner:
Inclusion of temporal information into features for speech recognition. ICSLP 1996: 256-259 - [c10]Saeed Vaseghi, Ben P. Milner:
A comparitive analysis of channel-robust features and channel equalization methods for speech recognition. ICSLP 1996: 877-880 - [c9]Naomi Harte, Saeed Vaseghi, Ben P. Milner:
Dynamic features for segmental speech recognition. ICSLP 1996: 933-936 - 1995
- [c8]Saeed Vaseghi, Ben P. Milner:
Speech recognition in impulsive noise. ICASSP 1995: 437-440 - [c7]Ben P. Milner, Saeed Vaseghi:
An analysis of cepstral-time matrices for noise and channel robust speech recognition. EUROSPEECH 1995: 519-522 - 1994
- [c6]Saeed Vaseghi, Ben P. Milner, Jason J. Humphries:
Noisy speech recognition using cepstral-time features and spectral-time filters. ICASSP (2) 1994: 65-68 - [c5]Ben P. Milner, Saeed Vaseghi:
Speech modelling using cepstral-time feature matrices and hidden Markov models. ICASSP (1) 1994: 601-604 - 1993
- [c4]Saeed Vaseghi, Ben P. Milner:
Noisy speech recognition based on HMMs, Wiener filters and re-evaluation of most likely candidates. ICASSP (2) 1993: 103-106 - [c3]Saeed Vaseghi, P. N. Conner, Ben P. Milner:
Speech modelling using cepstral-time feature matrices. EUROSPEECH 1993: 591-594 - [c2]Saeed Vaseghi, Ben P. Milner:
Noise-adaptive hidden Markov models based on wiener filters. EUROSPEECH 1993: 1023-1026 - 1992
- [c1]Saeed Vaseghi, Ben P. Milner:
Speech recognition in noisy environments. ICSLP 1992: 1487-1490
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-10-07 22:17 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint