Nothing Special   »   [go: up one dir, main page]

Llerena-Aguilar et al., 2017 - Google Patents

A new mixing matrix estimation method based on the geometrical analysis of the sound separation problem

Llerena-Aguilar et al., 2017

Document ID
4970286810850320714
Author
Llerena-Aguilar C
Gil-Pita R
Utrilla-Manso M
Rosa-Zurera M
Publication year
Publication venue
Signal Processing

External Links

Snippet

Nowadays, some of the most successful sound source separation methods are based on the assumption of sparse sources. A large number of those separation solutions consist of two parts: the mixing matrix estimation and the separation stages. Concerning the first part, many …
Continue reading at www.sciencedirect.com (other versions)

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • G10L2021/02161Number of inputs available containing the signal or the noise to be suppressed
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S3/00Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received
    • G01S3/80Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received using ultrasonic, sonic or infrasonic waves
    • G01S3/802Systems for determining direction or deviation from predetermined direction
    • G01S3/808Systems for determining direction or deviation from predetermined direction using transducers spaced apart and measuring phase or time difference between signals therefrom, i.e. path-difference systems
    • G01S3/8083Systems for determining direction or deviation from predetermined direction using transducers spaced apart and measuring phase or time difference between signals therefrom, i.e. path-difference systems determining direction of source
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/301Automatic calibration of stereophonic sound system, e.g. with test microphone
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/15Aspects of sound capture and related signal processing for recording or reproduction

Similar Documents

Publication Publication Date Title
Mandel et al. Model-based expectation-maximization source separation and localization
Nadiri et al. Localization of multiple speakers under high reverberation using a spherical microphone array and the direct-path dominance test
Mandel et al. An EM algorithm for localizing multiple sound sources in reverberant environments
Nikunen et al. Direction of arrival based spatial covariance model for blind sound source separation
CN109839612A (en) Sounnd source direction estimation method based on time-frequency masking and deep neural network
Georganti et al. Sound source distance estimation in rooms based on statistical properties of binaural signals
Delikaris-Manias et al. 3D localization of multiple audio sources utilizing 2D DOA histograms
Dadvar et al. Robust binaural speech separation in adverse conditions based on deep neural network with modified spatial features and training target
Cobos et al. Two-microphone multi-speaker localization based on a Laplacian mixture model
Lathoud et al. A sector-based, frequency-domain approach to detection and localization of multiple speakers
Mack et al. Signal-aware direction-of-arrival estimation using attention mechanisms
Aarabi et al. Robust sound localization using conditional time–frequency histograms
Thiergart et al. Localization of sound sources in reverberant environments based on directional audio coding parameters
Llerena-Aguilar et al. A new mixing matrix estimation method based on the geometrical analysis of the sound separation problem
Ding et al. DOA estimation of multiple speech sources by selecting reliable local sound intensity estimates
Cai et al. Accelerated steered response power method for sound source localization using orthogonal linear array
Riaz et al. Adaptive blind moving source separation based on intensity vector statistics
Georganti et al. Extracting sound-source-distance information from binaural signals
Sun et al. Indoor multiple sound source localization using a novel data selection scheme
Berdugo et al. Speakers’ direction finding using estimated time delays in the frequency domain
Schwartz et al. Array Configuration Mismatch in Deep DOA Estimation: Towards Robust Training
Gul et al. Preserving the beamforming effect for spatial cue-based pseudo-binaural dereverberation of a single source
Mandel et al. A probability model for interaural phase difference
Dang et al. An iteratively reweighted steered response power approach to multisource localization using a distributed microphone network
Gburrek et al. On source-microphone distance estimation using convolutional recurrent neural networks