KL quantifies the distance between two probability distributions: Let and be two probability distributions of a discrete random variable :

By convention :

  • Properties:

Kullback-Leibler Divergence : Examples

For a binary random variable with range , assume two distributions and are defined as :

KL Divergences :