STAT 400

Fri. February 7th, 2019


Independence (cont.)

To recap: two events AA and BB are independent if and only if P(AB)=P(A)P(A\mid B)=P(A), and dependent otherwise.

Important: Just because two events are dependent does not mean that one event was caused by the other; that is to say, dependence is not the same as causality. In statistics, we only study the correlation between events.


Properties of Independence

  1. Independence is symmetric: P(AB)=P(A)P(BA)=P(B)P(A\mid B)=P(A)\leftrightarrow P(B\mid A)=P(B).
  2. If AA and BB are independent, then AA' and BB are also independent:
    P(AB)=P(BA)P(B)=P(B)P(BA)P(B)=1P(BA)P(B)=1P(AB)=1P(A)=P(A).\begin{aligned} P(A'\mid B)&=\frac{P(B\cap A')}{P(B)}\\&=\frac{P(B)-P(B\cap A)}{P(B)} \\&=1-\frac{P(B\cap A)}{P(B)}\\&=1-P(A\mid B)\\&=1-P(A)=P(A'). \end{aligned}
  3. AA and BB are independent if and only if P(AB)=P(A)P(B)P(A\cap B)=P(A)\cdot P(B).

Ex. Toss a coin ten times.
A={first 9 tosses are all H}B={10th toss is T}C={at least one T within the first 9 tosses}\begin{aligned} A&=\{\text{first 9 tosses are all H}\}\\ B&=\{\text{10th toss is T}\}\\ C&=\{\text{at least one T within the first 9 tosses}\} \end{aligned}

Intuitively:


Independence of more than two events

Definition. Given events A1,A2...AnA_1,A_2...A_n, we say they are mutually independent if, for every k=2,3,...nk=2,3,...n and every subset of indices i1,i2...iki_1,i_2...i_k, P(Ai1Ai2Ai3...Aik)=P(Ai1)P(Ai2)...P(Aik).P(A_{i_1}\cap A_{i_2}\cap A_{i_3}...\cap A_{i_k})=P(A_{i_1})\cdot P(A_{i_2})...P(A_{i_k}).

For instance, if we want to determine if A1,A2,A3A_1, A_2, A_3 are mutually independent, then all of the following equations must hold:

The number of equations needed to check grows exponentially, so you probably won’t be asked to check more than three events.


Ex. Roll a die twice. A1={first roll is 3}A2={second roll is 4}A3={sum of the two rolls is 7}\begin{aligned} A_1&=\{\text{first roll is 3}\}\\ A_2&=\{\text{second roll is 4}\}\\ A_3&=\{\text{sum of the two rolls is 7}\} \end{aligned} Are these events mutually independent?

Because all of the k=2k=2 cases are true, we call these events pairwise independent. However, the k=3k=3 case is false, so they are not mutually independent.


Random Variables

We have already studied the probability of random outcomes occurring quantitatively, but we also want to be able to quantitatively study the outcomes themselves. Often these already have a numerical value associated with them (like a die roll), but sometimes they don’t (like a coin flip). In these cases we can create our own mapping to numerical values, for instance: {H,T}{1,0}.\{H,T\}\rightarrow \{1,0\}.

Definition. Given a sample space of some experiment, a random variable is a rule that associates a value with each outcome in SS; that is, a random variable XX is a mapping from SS to R\R. X:SR.X: S \rightarrow \R.