Evaluating Bregman Divergences for Probability Learning from Crowd

by   F. A. Mena, et al.
Universidad Técnica Federico Santa María

The crowdsourcing scenarios are a good example of having a probability distribution over some categories showing what the people in a global perspective thinks. Learn a predictive model of this probability distribution can be of much more valuable that learn only a discriminative model that gives the most likely category of the data. Here we present differents models that adapts having probability distribution as target to train a machine learning model. We focus on the Bregman divergences framework to used as objective function to minimize. The results show that special care must be taken when build a objective function and consider a equal optimization on neural network in Keras framework.


page 1

page 2

page 3

page 4


SPOT: A framework for selection of prototypes using optimal transport

In this work, we develop an optimal transport (OT) based framework to se...

Distributionally Robust Bayesian Quadrature Optimization

Bayesian quadrature optimization (BQO) maximizes the expectation of an e...

Assessing the Performance of Deep Learning Algorithms for Newsvendor Problem

In retailer management, the Newsvendor problem has widely attracted atte...

Probability Distribution of Hypervolume Improvement in Bi-objective Bayesian Optimization

This work provides the exact expression of the probability distribution ...

Shallow Neural Network can Perfectly Classify an Object following Separable Probability Distribution

Guiding the design of neural networks is of great importance to save eno...

Probability Distribution on Rooted Trees

The hierarchical and recursive expressive capability of rooted trees is ...

Error Analysis of Surrogate Models Constructed through Operations on Sub-models

Model-based methods are popular in derivative-free optimization (DFO). I...

Please sign up or login with your details

Forgot password? Click here to reset