Top Qs
Timeline
Chat
Perspective
Speaker diarisation
Partitioning a stream of human speech by identity of speaker From Wikipedia, the free encyclopedia
Remove ads
Speaker diarisation (or diarization) is the process of partitioning an audio stream containing human speech into homogeneous segments according to the identity of each speaker.[1] It can enhance the readability of an automatic speech transcription by structuring the audio stream into speaker turns and, when used together with speaker recognition systems, by providing the speaker’s true identity.[2] It is used to answer the question "who spoke when?"[3] Speaker diarisation is a combination of speaker segmentation and speaker clustering. The first aims at finding speaker change points in an audio stream. The second aims at grouping together speech segments on the basis of speaker characteristics.
This article needs additional citations for verification. (January 2012) |
With the increasing number of broadcasts, meeting recordings and voice mail collected every year, speaker diarisation has received much attention by the speech community, as is manifested by the specific evaluations devoted to it under the auspices of the National Institute of Standards and Technology for telephone speech, broadcast news and meetings.[4] A leading list tracker of speaker diarization research can be found at Quan Wang's github repo.[5]
Remove ads
Main types of diarisation systems
In speaker diarisation, one of the most popular methods is to use a Gaussian mixture model to model each of the speakers, and assign the corresponding frames for each speaker with the help of a Hidden Markov Model. There are two main kinds of clustering strategies. Bottom-up algorithms are the most popular, and they start by splitting the full audio content in a succession of clusters and progressively try to merge the redundant clusters in order to reach a situation where each cluster corresponds to a real speaker. The second, top-down algorithms, start with a single cluster for all the audio data and try to split them iteratively until reaching a number of clusters equal to the number of speakers.
More recently, speaker diarisation is performed via neural networks leveraging large-scale GPU computing and methodological developments in deep learning.[6]
Remove ads
References
Bibliography
Wikiwand - on
Seamless Wikipedia browsing. On steroids.
Remove ads