Representation Disentaglement via Regularization by Identification

02/28/2023
by   Juan Castorena, et al.
0

This work focuses on the problem of learning disentangled representations from observational data. Given observations 𝐱^(i) for i=1,...,N drawn from p(𝐱|𝐲) with generative variables 𝐲 admitting the distribution factorization p(𝐲) = ∏_c p(𝐲_c ) we ask whether learning disentangled representations matching the space of observations with identification guarantees on the posterior p(𝐳| 𝐱, 𝐲̂_c) for each c, is plausible. We argue modern deep representation learning models are ill-posed with collider bias behaviour; a source of bias producing entanglement between generating variables. Under the rubric of causality, we show this issue can be explained and reconciled under the condition of identifiability; attainable under supervision or a weak-form of it. For this, we propose regularization by identification (ReI), a regularization framework defined by the identification of the causal queries involved in the learning problem. Empirical evidence shows that enforcing ReI in a variational framework results in disentangled representations equipped with generalization capabilities to out-of-distribution examples and that aligns nicely with the true expected effect between generating variables and measurement apparatus.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset