Mutual Information Decay Curves and Hyper-Parameter Grid Search Design for Recurrent Neural Architectures

12/08/2020
by   Abhijit Mahalunkar, et al.
0

We present an approach to design the grid searches for hyper-parameter optimization for recurrent neural architectures. The basis for this approach is the use of mutual information to analyze long distance dependencies (LDDs) within a dataset. We also report a set of experiments that demonstrate how using this approach, we obtain state-of-the-art results for DilatedRNNs across a range of benchmark datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/18/2018

Dynamic learning rate using Mutual Information

This paper demonstrates dynamic hyper-parameter setting, for deep neural...
research
10/06/2018

Understanding Recurrent Neural Architectures by Analyzing and Synthesizing Long Distance Dependencies in Benchmark Sequential Datasets

At present, the state-of-the-art computational models across a range of ...
research
06/10/2020

On the Maximum Mutual Information Capacity of Neural Architectures

We derive the closed-form expression of the maximum mutual information -...
research
01/30/2023

Quantifying and maximizing the information flux in recurrent neural networks

Free-running Recurrent Neural Networks (RNNs), especially probabilistic ...
research
07/13/2019

Multi-Element Long Distance Dependencies: Using SPk Languages to Explore the Characteristics of Long-Distance Dependencies

In order to successfully model Long Distance Dependencies (LDDs) it is n...
research
08/15/2018

Using Regular Languages to Explore the Representational Capacity of Recurrent Neural Architectures

The presence of Long Distance Dependencies (LDDs) in sequential data pos...
research
05/10/2019

Mutual Information Scaling and Expressive Power of Sequence Models

Sequence models assign probabilities to variable-length sequences such a...

Please sign up or login with your details

Forgot password? Click here to reset