Gannot et al., 2017 - Google Patents
A consolidated perspective on multimicrophone speech enhancement and source separationGannot et al., 2017
View PDF- Document ID
- 7085213525392481648
- Author
- Gannot S
- Vincent E
- Markovich-Golan S
- Ozerov A
- Publication year
- Publication venue
- IEEE/ACM Transactions on Audio, Speech, and Language Processing
External Links
Snippet
Speech enhancement and separation are core problems in audio signal processing, with commercial applications in devices as diverse as mobile phones, conference call systems, hands-free systems, or hearing aids. In addition, they are crucial preprocessing steps for …
- 238000000926 separation method 0 title abstract description 41
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L21/0216—Noise filtering characterised by the method used for estimating noise
- G10L2021/02161—Number of inputs available containing the signal or the noise to be suppressed
- G10L2021/02166—Microphone arrays; Beamforming
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/065—Adaptation
- G10L15/07—Adaptation to the speaker
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signal, using source filter models or psychoacoustic analysis
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/005—Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Gannot et al. | A consolidated perspective on multimicrophone speech enhancement and source separation | |
Wang et al. | Deep learning based target cancellation for speech dereverberation | |
Wang et al. | Complex spectral mapping for single-and multi-channel speech enhancement and robust ASR | |
Cobos et al. | Frequency-sliding generalized cross-correlation: A sub-band time delay estimation approach | |
Kumatani et al. | Microphone array processing for distant speech recognition: From close-talking microphones to far-field sensors | |
US7366662B2 (en) | Separation of target acoustic signals in a multi-transducer arrangement | |
Perotin et al. | Multichannel speech separation with recurrent neural networks from high-order ambisonics recordings | |
Taseska et al. | Informed spatial filtering for sound extraction using distributed microphone arrays | |
Wang et al. | Noise power spectral density estimation using MaxNSR blocking matrix | |
Schwartz et al. | An expectation-maximization algorithm for multimicrophone speech dereverberation and noise reduction with coherence matrix estimation | |
Koldovský et al. | Spatial source subtraction based on incomplete measurements of relative transfer function | |
Nakatani et al. | Dominance based integration of spatial and spectral features for speech enhancement | |
Kumatani et al. | Beamforming with a maximum negentropy criterion | |
Nesta et al. | A flexible spatial blind source extraction framework for robust speech recognition in noisy environments | |
Kolossa et al. | CHiME challenge: Approaches to robustness using beamforming and uncertainty-of-observation techniques | |
Habets et al. | Dereverberation | |
Malik et al. | A Bayesian framework for blind adaptive beamforming | |
Šarić et al. | Supervised speech separation combined with adaptive beamforming | |
Kühne et al. | A new evidence model for missing data speech recognition with applications in reverberant multi-source environments | |
Markovich‐Golan et al. | Spatial filtering | |
Yu | Post-filter optimization for multichannel automotive speech enhancement | |
Kindt et al. | Improved separation of closely-spaced speakers by exploiting auxiliary direction of arrival information within a u-net architecture | |
Taghia et al. | Dual-channel noise reduction based on a mixture of circular-symmetric complex Gaussians on unit hypersphere | |
Nakatani et al. | Simultaneous denoising, dereverberation, and source separation using a unified convolutional beamformer | |
Adcock | Optimal filtering and speech recognition with microphone arrays |