Definition (Cross entropy)
Given two pmfs and on the same alphabet , the Cross-Entropy between and , denoted as is
Remark
Cross-Entropy is a quantity intimately connected to divergence:
Where divergence essentially tells us the difference in entropy between using to encode and using to encode . Cross-Entropy tells us the total number of expected bits (or entropy) needed to when symbols are generated from but is used to encode them.
Proposition (Entropy vs Cross-Entropy)
Since , then by the above remark we have which implies