• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 97
Next 10 →

An Adaptive Close-Talking Microphone Array

by Heinz Teutsch, Gary W. Elko - IEEE Workshop on Applications of Signal Processing to Audio and Acoustics , 2001
"... Close-talking differential microphone arrays (CTMAs) are useful in situations where the background noise level is very high because they inherently suppress farfield noise while emphasizing desired nearfield signals. One problem, however, is that the array has to be placed as close to the desired so ..."
Abstract - Cited by 4 (1 self) - Add to MetaCart
Close-talking differential microphone arrays (CTMAs) are useful in situations where the background noise level is very high because they inherently suppress farfield noise while emphasizing desired nearfield signals. One problem, however, is that the array has to be placed as close to the desired

Detection of Laughter-in-Interaction in Multichannel Close-Talk Microphone Recordings of Meetings

by Kornel Laskowski, Tanja Schultz
"... Abstract. Laughter is a key element of human-human interaction, occurring surprisingly frequently in multi-party conversation. In meetings, laughter accounts for almost 10 % of vocalization effort by time, and is known to be relevant for topic segmentation and the automatic characterization of affec ..."
Abstract - Cited by 7 (3 self) - Add to MetaCart
Abstract. Laughter is a key element of human-human interaction, occurring surprisingly frequently in multi-party conversation. In meetings, laughter accounts for almost 10 % of vocalization effort by time, and is known to be relevant for topic segmentation and the automatic characterization of affect. We present a system for the detection of laughter, and its attribution to specific participants, which relies on simultaneously decoding the vocal activity of all participants given multi-channel recordings. The proposed framework allows us to disambiguate laughter and speech not only acoustically, but also by constraining the number of simultaneous speakers and the number of simultaneous laughers independently, since participants tend to take turns speaking but laugh together. We present experiments on 57 hours of meeting data, containing almost 11000 unique instances of laughter. 1

A SUPERVISED FACTORIAL ACOUSTIC MODEL FOR SIMULTANEOUS MULTIPARTICIPANT VOCAL ACTIVITY DETECTION IN CLOSE-TALK MICROPHONE RECORDINGS OF MEETINGS

by Kornel Laskowski, Tanja Schultz
"... Vocal activity detection in close-talk microphone recordings of multiparty conversation continues to pose problems for meeting recognition systems, as evidenced by a 2-3 % absolute gap in word error rates achieved with automatic and manual segmentations. State-of-the-art segmentation systems in this ..."
Abstract - Add to MetaCart
Vocal activity detection in close-talk microphone recordings of multiparty conversation continues to pose problems for meeting recognition systems, as evidenced by a 2-3 % absolute gap in word error rates achieved with automatic and manual segmentations. State-of-the-art segmentation systems

Microphone array processing for distant speech recognition: From close-talking microphones to far-field sensors,”

by Kenichi Kumatani , Takayuki Arakawa , Kazumasa Yamamoto , John Mcdonough , Raj ¶ Bhiksha , Rita Singh , Ivan Tashev - IEEE Signal Processing Magazine, , 2012
"... Abstract-Distant speech recognition (DSR) holds out the promise of providing a natural human computer interface in that it enables verbal interactions with computers without the necessity of donning intrusive body-or head-mounted devices. Recognizing distant speech robustly, however, remains a chal ..."
Abstract - Cited by 16 (4 self) - Add to MetaCart
challenge. This paper provides a overview of DSR systems based on microphone arrays. In particular, we present recent work on acoustic beamforming for DSR, along with experimental results verifying the effectiveness of the various algorithms described here; beginning from a word error rate (WER) of 14

Online Speaker Change Detection by Combining BIC with Microphone Array

by Joerg Schmalenstroeer, Reinhold Haeb-umbach
"... In this paper we consider the problem of detecting speaker changes in audio signals recorded by distant microphones. It is shown that the possibility to exploit the spatial separation of speakers more than makes up the degradation in detection accuracy due to the increased source-to-sensor distance ..."
Abstract - Cited by 6 (3 self) - Add to MetaCart
compared to close-talking microphones. Speaker direction information is derived from the filter coefficients of an adaptive Filter-and-Sum Beamformer and is combined with BIC analysis. The experimental results reveal significant improvements compared to BIC-only change detection, be it with the distant

FAU IISAH Corpus — A German Speech Database Consisting of Human-Machine and Human-Human Interaction Acquired by Close-Talking and Far-Distance Microphones

by Werner Spiegl, Korbinian Riedhammer, Stefan Steidl, Elmar Nöth
"... In this paper the FAU IISAH corpus and its recording conditions are described: a new speech database consisting of human-machine and human-human interaction recordings. Beside close-talking microphones for the best possible audio quality of the recorded speech, far-distance microphones were used to ..."
Abstract - Add to MetaCart
In this paper the FAU IISAH corpus and its recording conditions are described: a new speech database consisting of human-machine and human-human interaction recordings. Beside close-talking microphones for the best possible audio quality of the recorded speech, far-distance microphones were used

Acoustical and Environmental Robustness in Automatic Speech Recognition

by Alejandro Acero , 1990
"... This dissertation describes a number of algorithms developed to increase the robustness of automatic speech recognition systems with respect to changes in the environment. These algorithms attempt to improve the recognition accuracy of speech recognition systems when they are trained and tested in d ..."
Abstract - Cited by 214 (13 self) - Add to MetaCart
in different acoustical environments, and when a desk-top microphone (rather than a close-talking microphone) is used for speech input. Without such processing, mismatches between training and testing conditions produce an unacceptable degradation in recognition accuracy. Two kinds of

The ICSI Meeting Corpus: Close-talking and Far-field, Multi-channel Transcriptions for Speech and Language Researchers

by Jane A. Edwards
"... The recently-completed ICSI Meeting Corpus is available through the LDC. It consists of audio and transcripts of 75 research meetings, ranging in size from 3 to 10 people, with an average of 6 people. The meetings were recorded by means of both closetalking (headset or lapel) microphones and far-fie ..."
Abstract - Add to MetaCart
-field (table-top) microphones. The close-talking microphones enable separation of each person’s audible activities from those of every other participant. The far-field microphones provide a view of the meeting as a whole. The transcripts preserve words and other communicative phenomena, displayed in musical

Using Energy Difference for Speech Separation of Dual-microphone Close-talk System

by Yi Jiang, Ming Jiang, Yuanyuan Zu, Hong Zhou, Zhenming Feng , 2013
"... Abstract: Using the computational auditory scene analysis (CASA) as a framework, a novel speech separation approach based on dual-microphone energy difference (DMED) is proposed for close-talk system. The energy levels of the two microphones are calculated in time-frequency (T-F) units. The DMEDs ar ..."
Abstract - Add to MetaCart
Abstract: Using the computational auditory scene analysis (CASA) as a framework, a novel speech separation approach based on dual-microphone energy difference (DMED) is proposed for close-talk system. The energy levels of the two microphones are calculated in time-frequency (T-F) units. The DMEDs

Microphone Array Speech Recognition : Experiments on Overlapping Speech in Meetings

by Darren C. Moore, Iain A. Mccowan - in Proceedings of the International Conference on Acoustics, Speech and Signal Processing , 2003
"... This paper investigates the use of microphone arrays to acquire and recognise speech in meetings. Meetings pose several interesting problems for speech processing, as they consist of multiple competing speakers within a small space, typically around a table. Due to their ability to provide hands-fre ..."
Abstract - Cited by 35 (5 self) - Add to MetaCart
-free acquisition and directional discrimination, microphone arrays present a potential alternative to close-talking microphones in such an application. We first propose an appropriate microphone array geometry and improved processing technique for this scenario, paying particular attention to speaker separation
Next 10 →
Results 1 - 10 of 97
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University