Definition

Assume that two Probability Distributions and are given. Then the Kullback-Leibler divergence between and is defined as
Kullback-Leibler divergence measures how different two distributions are.
It also can be expressed as a difference between the cross entropy (difference between distributions and )and entropy (inherent uncertainty of ).
Facts
Let be a sequence of distributions. Then, The convergence of the KL-Divergence to zero implies that the JS-Divergence also converges to zero. The convergence of the JS-Divergence to zero is equivalent to the convergence of the Total Variation Distance to zero. The convergence of the Total Variation Distance to zero implies that the Wasserstein Distance also converges to zero. The convergence of the Wasserstein Distance to zero is equivalent to the Convergence in Distribution of the sequence.