1

We know that for discrete variables \begin{equation} D(p(x),q(x))=\mathbb{E}_{p}\left(\log\frac{p(x)}{q(x)}\right) \end{equation} where $p(x)$ and $q(x)$ are probability mass functions. Can this be extended to \begin{equation} D(p(x,y,z),q(x,y,z))=\mathbb{E}_{p}\left(\log\frac{p(x,y,z)}{q(x,y,z)}\right) \end{equation} ? where now $p(x,y,z)$ and $q(x,y,z)$ are joint probability mass functions. Is the notation correct?

Carl
  • 1,196

1 Answers1

1

Yes, that's correct. The way you've written KL divergence using expectations is valid for any pair of distributions defined on the same space (not limited to discrete values; also works for reals, vectors, etc.). Notation looks fine to me, as long as you specify that (⋅,⋅) means KL divergence. Everyone writes things a little bit differently. – user20160


I've created this answer from a comment so that this question is answered. Better to have a short answer than no answer at all.

Sycorax
  • 90,934