denoted as DKL(P∥Q)
The statistical distance between a model probability distribution Q difference from a true probability distribution P:
DKL(P∥Q)=x∈X∑P(x)log(Q(x)P(x))
Alternative form 1:
KL(p∥q)=Ex∼p(logq(x)p(x))=∫xP(x)logq(x)p(x)dx
For relative entropy if ∀x>0,Q(x)=0⟹P(x)=0 absolute continuity
For distribution P and Q of a continuous random variable, then KL divergence is:
DKL(P∥Q)=∫−∞+∞p(x)logq(x)p(x)dx
where p and q denote probability densities of P and Q