Speaker Diarization: Using Recurrent Neural Networks

06/10/2020
by   Vishal Sharma, et al.
0

Speaker Diarization is the problem of separating speakers in an audio. There could be any number of speakers and final result should state when speaker starts and ends. In this project, we analyze given audio file with 2 channels and 2 speakers (on separate channel). We train Neural Network for learning when a person is speaking. We use different type of Neural Networks specifically, Single Layer Perceptron (SLP), Multi Layer Perceptron (MLP), Recurrent Neural Network (RNN) and Convolution Neural Network (CNN) we achieve ∼92% of accuracy with RNN. The code for this project is available at https://github.com/vishalshar/SpeakerDiarization_RNN_CNN_LSTM

READ FULL TEXT

page 7

page 8

research
11/01/2017

Performance Evaluation of Channel Decoding With Deep Neural Networks

With the demand of high data rate and low latency in fifth generation (5...
research
10/10/2018

Fully Supervised Speaker Diarization

In this paper, we propose a fully supervised speaker diarization approac...
research
03/23/2018

Convolutional vs. Recurrent Neural Networks for Audio Source Separation

Recent work has shown that recurrent neural networks can be trained to s...
research
03/28/2022

Training speaker recognition systems with limited data

This work considers training neural networks for speaker recognition wit...
research
01/06/2021

Multichannel CRNN for Speaker Counting: an Analysis of Performance

Speaker counting is the task of estimating the number of people that are...
research
04/07/2021

Three-class Overlapped Speech Detection using a Convolutional Recurrent Neural Network

In this work, we propose an overlapped speech detection system trained a...
research
04/28/2019

Application of Autoencoder-Assisted Recurrent Neural Networks to Prevent Cases of Sudden Infant Death Syndrome

This project develops and trains a Recurrent Neural Network (RNN) that m...

Please sign up or login with your details

Forgot password? Click here to reset