On Learning Associations of Faces and Voices

05/15/2018
by   Changil Kim, et al.
0

In this paper, we study the associations between human faces and voices. Audiovisual integration (AVI), specifically the integration of facial and vocal information is a well-researched area in neuroscience. It is well known that the overlapping information between the two modalities plays a significant role in perceptual tasks such as speaker identification. Through an online study, we confirm previous findings that people can associate faces with corresponding voices and vice versa with greater than chance accuracy. We show that machines can learn such associations and use the learned information to identify matching faces and voices, with close to human performance. We analyze our findings statistically and evaluate our learned representations.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset