Loss Adapted Plasticity in Deep Neural Networks to Learn from Data with Unreliable Sources

12/06/2022
by   Alexander Capstick, et al.
0

When data is streaming from multiple sources, conventional training methods update model weights often assuming the same level of reliability for each source; that is: a model does not consider data quality of each source during training. In many applications, sources can have varied levels of noise or corruption that has negative effects on the learning of a robust deep learning model. A key issue is that the quality of data or labels for individual sources is often not available during training and could vary over time. Our solution to this problem is to consider the mistakes made while training on data originating from sources and utilise this to create a perceived data quality for each source. This paper demonstrates a straight-forward and novel technique that can be applied to any gradient descent optimiser: Update model weights as a function of the perceived reliability of data sources within a wider data set. The algorithm controls the plasticity of a given model to weight updates based on the history of losses from individual data sources. We show that applying this technique can significantly improve model performance when trained on a mixture of reliable and unreliable data sources, and maintain performance when models are trained on data sources that are all considered reliable. All code to reproduce this work's experiments and implement the algorithm in the reader's own models is made available.

READ FULL TEXT
research
01/01/2023

An Adaptive Kernel Approach to Federated Learning of Heterogeneous Causal Effects

We propose a new causal inference framework to learn causal effects from...
research
03/27/2023

Generalisability of deep learning-based early warning in the intensive care unit: a retrospective empirical evaluation

Deep learning (DL) can aid doctors in detecting worsening patient states...
research
04/13/2022

Achieving Representative Data via Convex Hull Feasibility Sampling Algorithms

Sampling biases in training data are a major source of algorithmic biase...
research
10/28/2020

Online feature selection for rapid, low-overhead learning in networked systems

Data-driven functions for operation and management often require measure...
research
11/02/2022

Thunderstorm nowcasting with deep learning: a multi-hazard data fusion model

Predictions of thunderstorm-related hazards are needed in several sector...
research
02/11/2021

Rethinking Representations in P C Actuarial Science with Deep Neural Networks

Insurance companies gather a growing variety of data for use in the insu...
research
02/28/2023

Gradient-Boosted Based Structured and Unstructured Learning

We propose two frameworks to deal with problem settings in which both st...

Please sign up or login with your details

Forgot password? Click here to reset