Joint Entropy
This online calculator calculates joint entropy of two discrete random variables given a joint distribution table (X, Y) ~ p
Joint entropy is a measure of "the uncertainty" associated with a set of variables.
In order to calculate the joint entropy, you should enter the joint distribution matrix where the cell value for any i row and j column represents the probability of the outcome,
. You can find the joint entropy formula below the calculator.
Joint Entropy Formula
The joint Shannon entropy (in bits) of two discrete random variables and
with images
and
is defined as:
where and
are particular values of
and
, respectively,
is the joint probability of these values occurring together, and
is defined to be 0 if
1
URL copied to clipboard
Similar calculators
PLANETCALC, Joint Entropy
Comments