Invertible Low-Divergence Coding

10/20/2020
by   Patrick Schulte, et al.
0

Several applications in communication, control, and learning require approximating target distributions to within small informational divergence (I-divergence). The additional requirement of invertibility usually leads to using encoders that are one-to-one mappings, also known as distribution matchers. However, even the best one-to-one encoders have I-divergences that grow logarithmically with the block length in general. To improve performance, an encoder is proposed that has an invertible one-to-many mapping and a low-rate resolution code. Two algorithms are developed to design the mapping by assigning strings in either a most-likely first or least-likely first order. Both algorithms give information rates approaching the entropy of the target distribution with exponentially decreasing I-divergence and with vanishing resolution rate in the block length.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro