Relaxed Weight Sharing: Effectively Modeling Time-Varying Relationships in Clinical Time-Series

06/07/2019
by   Jeeheh Oh, et al.
0

Recurrent neural networks (RNNs) are commonly applied to clinical time-series data with the goal of learning patient risk stratification models. Their effectiveness is due, in part, to their use of parameter sharing over time (i.e., cells are repeated hence the name recurrent). We hypothesize, however, that this trait also contributes to the increased difficulty such models have with learning relationships that change over time. Conditional shift, i.e., changes in the relationship between the input X and the output y, arises if the risk factors for the event of interest change over the course of a patient admission. While in theory, RNNs and gated RNNs (e.g., LSTMs) in particular should be capable of learning time-varying relationships, when training data are limited, such models often fail to accurately capture these dynamics. We illustrate the advantages and disadvantages of complete weight sharing (RNNs) by comparing an LSTM with shared parameters to a sequential architecture with time-varying parameters on three clinically-relevant prediction tasks: acute respiratory failure (ARF), shock, and in-hospital mortality. In experiments using synthetic data, we demonstrate how weight sharing in LSTMs leads to worse performance in the presence of conditional shift. To improve upon the dichotomy between complete weight sharing vs. no weight sharing, we propose a novel RNN formulation based on a mixture model in which we relax weight sharing over time. The proposed method outperforms standard LSTMs and other state-of-the-art baselines across all tasks. In settings with limited data, relaxed weight sharing can lead to improved patient risk stratification performance.

READ FULL TEXT
research
02/23/2021

Model-Attentive Ensemble Learning for Sequence Modeling

Medical time-series datasets have unique characteristics that make predi...
research
11/29/2018

Leveraging Clinical Time-Series Data for Prediction: A Cautionary Tale

In healthcare, patient risk stratification models are often learned usin...
research
04/09/2021

DeepSITH: Efficient Learning via Decomposition of What and When Across Time Scales

Extracting temporal relationships over a range of scales is a hallmark o...
research
10/08/2020

A Fully Tensorized Recurrent Neural Network

Recurrent neural networks (RNNs) are powerful tools for sequential model...
research
10/04/2022

Data drift correction via time-varying importance weight estimator

Real-world deployment of machine learning models is challenging when dat...
research
04/09/2020

Industrial Forecasting with Exponentially Smoothed Recurrent Neural Networks

Industrial forecasting has entered an era of unprecedented growth in the...
research
11/08/2017

Learning Credible Models

In many settings, it is important that a model be capable of providing r...

Please sign up or login with your details

Forgot password? Click here to reset