Multi-task Recurrent Model for Speech and Speaker Recognition

03/31/2016
by   Zhiyuan Tang, et al.
0

Although highly correlated, speech and speaker recognition have been regarded as two independent tasks and studied by two communities. This is certainly not the way that people behave: we decipher both speech content and speaker traits at the same time. This paper presents a unified model to perform speech and speaker recognition simultaneously and altogether. The model is based on a unified neural network where the output of one task is fed to the input of the other, leading to a multi-task recurrent network. Experiments show that the joint model outperforms the task-specific models on both the two tasks.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset