Results 1 - 10
of
97
An Adaptive Close-Talking Microphone Array
- IEEE Workshop on Applications of Signal Processing to Audio and Acoustics
, 2001
"... Close-talking differential microphone arrays (CTMAs) are useful in situations where the background noise level is very high because they inherently suppress farfield noise while emphasizing desired nearfield signals. One problem, however, is that the array has to be placed as close to the desired so ..."
Abstract
-
Cited by 4 (1 self)
- Add to MetaCart
Close-talking differential microphone arrays (CTMAs) are useful in situations where the background noise level is very high because they inherently suppress farfield noise while emphasizing desired nearfield signals. One problem, however, is that the array has to be placed as close to the desired
Detection of Laughter-in-Interaction in Multichannel Close-Talk Microphone Recordings of Meetings
"... Abstract. Laughter is a key element of human-human interaction, occurring surprisingly frequently in multi-party conversation. In meetings, laughter accounts for almost 10 % of vocalization effort by time, and is known to be relevant for topic segmentation and the automatic characterization of affec ..."
Abstract
-
Cited by 7 (3 self)
- Add to MetaCart
Abstract. Laughter is a key element of human-human interaction, occurring surprisingly frequently in multi-party conversation. In meetings, laughter accounts for almost 10 % of vocalization effort by time, and is known to be relevant for topic segmentation and the automatic characterization of affect. We present a system for the detection of laughter, and its attribution to specific participants, which relies on simultaneously decoding the vocal activity of all participants given multi-channel recordings. The proposed framework allows us to disambiguate laughter and speech not only acoustically, but also by constraining the number of simultaneous speakers and the number of simultaneous laughers independently, since participants tend to take turns speaking but laugh together. We present experiments on 57 hours of meeting data, containing almost 11000 unique instances of laughter. 1
A SUPERVISED FACTORIAL ACOUSTIC MODEL FOR SIMULTANEOUS MULTIPARTICIPANT VOCAL ACTIVITY DETECTION IN CLOSE-TALK MICROPHONE RECORDINGS OF MEETINGS
"... Vocal activity detection in close-talk microphone recordings of multiparty conversation continues to pose problems for meeting recognition systems, as evidenced by a 2-3 % absolute gap in word error rates achieved with automatic and manual segmentations. State-of-the-art segmentation systems in this ..."
Abstract
- Add to MetaCart
Vocal activity detection in close-talk microphone recordings of multiparty conversation continues to pose problems for meeting recognition systems, as evidenced by a 2-3 % absolute gap in word error rates achieved with automatic and manual segmentations. State-of-the-art segmentation systems
Microphone array processing for distant speech recognition: From close-talking microphones to far-field sensors,”
- IEEE Signal Processing Magazine,
, 2012
"... Abstract-Distant speech recognition (DSR) holds out the promise of providing a natural human computer interface in that it enables verbal interactions with computers without the necessity of donning intrusive body-or head-mounted devices. Recognizing distant speech robustly, however, remains a chal ..."
Abstract
-
Cited by 16 (4 self)
- Add to MetaCart
challenge. This paper provides a overview of DSR systems based on microphone arrays. In particular, we present recent work on acoustic beamforming for DSR, along with experimental results verifying the effectiveness of the various algorithms described here; beginning from a word error rate (WER) of 14
Online Speaker Change Detection by Combining BIC with Microphone Array
"... In this paper we consider the problem of detecting speaker changes in audio signals recorded by distant microphones. It is shown that the possibility to exploit the spatial separation of speakers more than makes up the degradation in detection accuracy due to the increased source-to-sensor distance ..."
Abstract
-
Cited by 6 (3 self)
- Add to MetaCart
compared to close-talking microphones. Speaker direction information is derived from the filter coefficients of an adaptive Filter-and-Sum Beamformer and is combined with BIC analysis. The experimental results reveal significant improvements compared to BIC-only change detection, be it with the distant
FAU IISAH Corpus — A German Speech Database Consisting of Human-Machine and Human-Human Interaction Acquired by Close-Talking and Far-Distance Microphones
"... In this paper the FAU IISAH corpus and its recording conditions are described: a new speech database consisting of human-machine and human-human interaction recordings. Beside close-talking microphones for the best possible audio quality of the recorded speech, far-distance microphones were used to ..."
Abstract
- Add to MetaCart
In this paper the FAU IISAH corpus and its recording conditions are described: a new speech database consisting of human-machine and human-human interaction recordings. Beside close-talking microphones for the best possible audio quality of the recorded speech, far-distance microphones were used
Acoustical and Environmental Robustness in Automatic Speech Recognition
, 1990
"... This dissertation describes a number of algorithms developed to increase the robustness of automatic speech recognition systems with respect to changes in the environment. These algorithms attempt to improve the recognition accuracy of speech recognition systems when they are trained and tested in d ..."
Abstract
-
Cited by 214 (13 self)
- Add to MetaCart
in different acoustical environments, and when a desk-top microphone (rather than a close-talking microphone) is used for speech input. Without such processing, mismatches between training and testing conditions produce an unacceptable degradation in recognition accuracy. Two kinds of
The ICSI Meeting Corpus: Close-talking and Far-field, Multi-channel Transcriptions for Speech and Language Researchers
"... The recently-completed ICSI Meeting Corpus is available through the LDC. It consists of audio and transcripts of 75 research meetings, ranging in size from 3 to 10 people, with an average of 6 people. The meetings were recorded by means of both closetalking (headset or lapel) microphones and far-fie ..."
Abstract
- Add to MetaCart
-field (table-top) microphones. The close-talking microphones enable separation of each person’s audible activities from those of every other participant. The far-field microphones provide a view of the meeting as a whole. The transcripts preserve words and other communicative phenomena, displayed in musical
Using Energy Difference for Speech Separation of Dual-microphone Close-talk System
, 2013
"... Abstract: Using the computational auditory scene analysis (CASA) as a framework, a novel speech separation approach based on dual-microphone energy difference (DMED) is proposed for close-talk system. The energy levels of the two microphones are calculated in time-frequency (T-F) units. The DMEDs ar ..."
Abstract
- Add to MetaCart
Abstract: Using the computational auditory scene analysis (CASA) as a framework, a novel speech separation approach based on dual-microphone energy difference (DMED) is proposed for close-talk system. The energy levels of the two microphones are calculated in time-frequency (T-F) units. The DMEDs
Microphone Array Speech Recognition : Experiments on Overlapping Speech in Meetings
- in Proceedings of the International Conference on Acoustics, Speech and Signal Processing
, 2003
"... This paper investigates the use of microphone arrays to acquire and recognise speech in meetings. Meetings pose several interesting problems for speech processing, as they consist of multiple competing speakers within a small space, typically around a table. Due to their ability to provide hands-fre ..."
Abstract
-
Cited by 35 (5 self)
- Add to MetaCart
-free acquisition and directional discrimination, microphone arrays present a potential alternative to close-talking microphones in such an application. We first propose an appropriate microphone array geometry and improved processing technique for this scenario, paying particular attention to speaker separation
Results 1 - 10
of
97