Bias in Conversational Search: The Double-Edged Sword of the Personalized Knowledge Graph

by   Emma J. Gerritse, et al.

Conversational AI systems are being used in personal devices, providing users with highly personalized content. Personalized knowledge graphs (PKGs) are one of the recently proposed methods to store users' information in a structured form and tailor answers to their liking. Personalization, however, is prone to amplifying bias and contributing to the echo-chamber phenomenon. In this paper, we discuss different types of biases in conversational search systems, with the emphasis on the biases that are related to PKGs. We review existing definitions of bias in the literature: people bias, algorithm bias, and a combination of the two, and further propose different strategies for tackling these biases for conversational search systems. Finally, we discuss methods for measuring bias and evaluating user satisfaction.


page 1

page 2

page 3

page 4


Proceedings of the KG-BIAS Workshop 2020 at AKBC 2020

The KG-BIAS 2020 workshop touches on biases and how they surface in know...

BiasAsker: Measuring the Bias in Conversational AI System

Powered by advanced Artificial Intelligence (AI) techniques, conversatio...

Towards Automatic Bias Detection in Knowledge Graphs

With the recent surge in social applications relying on knowledge graphs...

A Survey on Bias in Visual Datasets

Computer Vision (CV) has achieved remarkable results, outperforming huma...

Uncovering Bias in Personal Informatics

Personal informatics (PI) systems, powered by smartphones and wearables,...

Personalized Detection of Cognitive Biases in Actions of Users from Their Logs: Anchoring and Recency Biases

Cognitive biases are mental shortcuts humans use in dealing with informa...

Personalized Emphasis Framing for Persuasive Message Generation

In this paper, we present a study on personalized emphasis framing which...

Please sign up or login with your details

Forgot password? Click here to reset