Kullback-Leibler Divergence and Akaike Information Criterion in General Hidden Markov Models

03/14/2023
by   Cheng-Der Fuh, et al.
0

To characterize the Kullback-Leibler divergence and Fisher information in general parametrized hidden Markov models, in this paper, we first show that the log likelihood and its derivatives can be represented as an additive functional of a Markovian iterated function system, and then provide explicit characterizations of these two quantities through this representation. Moreover, we show that Kullback-Leibler divergence can be locally approximated by a quadratic function determined by the Fisher information. Results relating to the Cramér-Rao lower bound and the Hájek-Le Cam local asymptotic minimax theorem are also given. As an application of our results, we provide a theoretical justification of using Akaike information criterion (AIC) model selection in general hidden Markov models. Last, we study three concrete models: a Gaussian vector autoregressive-moving average model of order (p,q), recurrent neural networks, and temporal restricted Boltzmann machine, to illustrate our theory.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset