Questions tagged [entropy]

A mathematical quantity designed to measure the amount of randomness of a random variable.

Entropy is a mathematical quantity designed to quantify the uncertainty about the occurrence of outcomes of a random variable. It is expressed as a function of the outcome probabilities of the random variable. Any measure for entropy must satisfy a few conditions:

  1. Continuity: The function must be continuous in all its arguments.
  2. Maximum: The function should have a maximum when all outcome are equally probable.
  3. Symmetry: The function must remain unchanged under a switch of arguments.

A commonly adopted measure is the Shannon entropy, $\mathrm{H}(p_1,p_2,\ldots,p_n)$ (when $p_1,p_2,\ldots p_n$ are the $n$ outcome probabilities of a random variable $X$). This measure is defined as follows:

$$\mathrm{H}(p_1,p_2,\ldots,p_n) = -\sum_{i=1}^n p_i \log p_i$$

698 questions
40
votes
3 answers

What does entropy tell us?

I am reading about entropy and am having a hard time conceptualizing what it means in the continuous case. The wiki page states the following: The probability distribution of the events, coupled with the information amount of every event, forms…
12
votes
3 answers

What is the significance of the log base being 2 in entropy?

What is the significance of the log base being 2 in entropy? What if we take e or 10 as the base?
7
votes
1 answer

Joint entropy of two random variables

The joint entropy is the amount of information we get when we observe X and Y at the same time, but what would happen if we don't observe them at the same time. For example, when i toss a coin, if i got tails i will only observe the variable X, but…
6
votes
1 answer

Is "Shannon entropy" used incorrectly in machine learning related literature?

When reading papers on machine learning, I have found that authors would often reference the "Shannon entropy". Curiously, often times the equation given would be: $$H(p) = -\sum\limits_{i = 1}^n p_i \ln(p_i)$$ For instance,…
Olórin
  • 724
5
votes
1 answer

Entropy of generalized distributions?

What's the entropy of the following generalized probability distributions? $P_1(x) = \delta(x)$ $P_2(x,y) = \delta(x+y)$, for $0\le x\le 1$, and $P_2(x,y)=0$ otherwise. Integrals of the type $-\int \delta(x) \ln\delta(x) \mathrm{d}x$ seem to diverge…
a06e
  • 4,410
  • 1
  • 22
  • 50
4
votes
2 answers

Entropy: Proving information gain formula: h(x) = -log p(x)

We consider a discrete random variable X, and we want to know how much information we receive every time we observe the value of this random variable. We qualify this measure of information transfer as h(x), a monotonically decreasing function of…
Mr Farad
  • 161
4
votes
1 answer

Understanding the entropy of a set

I have a limited statistics background, so I will try to be as specific as possible. I'm reading a paper on a method to fingerprint mobile devices using their accelerometer. To demonstrate that the method is robust, the author measures parameters…
4
votes
1 answer

Additivity of Shannon's entropy

E.T. Jaynes writes, in "Probability Theory: the Logic of Science" the following in order to motivate the derivation of the entropy, $H$: Suppose the robot perceives two alternatives, to which it assigns probabilities $p_1$ and $q := 1 - p_1$.…
3
votes
1 answer

Entropy (individuals equally high on all items vs. high on only few items)

I remember something from a stats course many years ago which might be helpful now. I want to distinguish between patients who show symptoms all over the board, vs. patients who have similarly high sum-scores, but due to a few very high symptoms.…
Torvon
  • 1,113
3
votes
1 answer

Significance of transfer entropy calculations

I have two time-series U and V (approximately 300 samples) having values between 0 and 5. I wanted to calculate the (first order) transfer entropy $I_{uv}$ (of U with the knowledge of V) and $I_{vu}$ (of V with the knowledge of U). Here are the…
AlexPof
  • 133
  • 4
3
votes
1 answer

Entropy stays the same with a larger distribution?

Based on the given definition of entropy, $H(P(X)) = -\sum_i P(x_i)log_2(P(x_i)$, it appears that if I have a distribution $P_1(x) = [\frac{1}{4},\frac{1}{4},\frac{1}{4},\frac{1}{4}]$ and another distribution $P_2(x) =…
Chet
  • 135
2
votes
0 answers

A distributional test based on entropy and self-information

Say that I have a real-valued discrete distribution $p(x)$ and $N$ samples, $x_1, \ldots, x_N$, and I want to test whether the samples came from the distribution without making any further assumptions whatsoever. Note that there are very few…
Paul
  • 10,920
2
votes
1 answer

Median entropy to observe evolution of system?

I am studying a dynamical system that takes as an initial condition a list. I want to analyze the evolution of Shannon's entropy in this system. I know the maximum entropy (50) and the minimum (0). Pure random conditions have almost maximum entropy,…
user2976
2
votes
0 answers

Entropy Balancing Did Not Create Balanced Means (R)

I am looking at how sex impacts judicial decision making and am trying to balance my dataset where my treatment variable is named "gender.judge" (0 for control group, males, and 1 for treatment group, females). I want to balance the data on these…
Liliana
  • 21
2
votes
1 answer

Negative Tsallis entropy?

I am building (many years later !) on this SO question: https://stackoverflow.com/questions/22461241/tsallis-entropy-for-continuous-variable-in-r I should make clear that I'm a stats newbie, exploring it out of curiosity, so forgive me if what…
Little Code
  • 151
  • 3
1
2 3