In information theory, the cross entropy between two probability distributions
, rather than the "true" distribution
The cross entropy for the distributions
For discrete
The situation for continuous distributions is analogous. We have to assume that
NB: The notation