Minimal model of permutation symmetry in unsupervised learning

04/30/2019
by   Tianqi Hou, et al.
0

Permutation of any two hidden units yields invariant properties in typical deep generative neural networks. This permutation symmetry plays an important role in understanding the computation performance of a broad class of neural networks with two or more hidden units. However, a theoretical study of the permutation symmetry is still lacking. Here, we propose a minimal model with only two hidden units in a restricted Boltzmann machine, which aims to address how the permutation symmetry affects the critical learning data size at which the concept-formation (or spontaneous symmetry breaking in physics language) starts, and moreover semi-rigorously prove a conjecture that the critical data size is independent of the number of hidden units once this number is finite. Remarkably, we find that the embedded correlation between two receptive fields of hidden units reduces the critical data size. In particular, the weakly-correlated receptive fields have the benefit of significantly reducing the minimal data size that triggers the transition, given less noisy data. Inspired by the theory, we also propose an efficient fully-distributed algorithm to infer the receptive fields of hidden units. Overall, our results demonstrate that the permutation symmetry is an interesting property that affects the critical data size for computation performances of related learning algorithms. All these effects can be analytically probed based on the minimal model, providing theoretical insights towards understanding unsupervised learning in a more general context.

READ FULL TEXT
research
11/06/2019

Statistical physics of unsupervised learning with prior knowledge in neural networks

Integrating sensory inputs with prior beliefs from past experiences in u...
research
07/03/2023

Learning permutation symmetries with gips in R

The study of hidden structures in data presents challenges in modern sta...
research
12/06/2016

Statistical mechanics of unsupervised feature learning in a restricted Boltzmann machine with binary synapses

Revealing hidden features in unlabeled data is called unsupervised featu...
research
03/17/2018

Replica Symmetry Breaking in Bipartite Spin Glasses and Neural Networks

Some interesting recent advances in the theoretical understanding of neu...
research
03/14/2018

Building Sparse Deep Feedforward Networks using Tree Receptive Fields

Sparse connectivity is an important factor behind the success of convolu...
research
08/21/2009

Computational Understanding and Manipulation of Symmetries

For natural and artificial systems with some symmetry structure, computa...

Please sign up or login with your details

Forgot password? Click here to reset