Conditional probability

From Wikipedia, the free encyclopedia

Jump to: navigation, search

This article defines some terms which characterize probability distributions of two or more variables.

Conditional probability is the probability of some event A, given the occurrence of some other event B. Conditional probability is written P(A|B), and is read "the probability of A, given B".

Joint probability is the probability of two events in conjunction. That is, it is the probability of both events together. The joint probability of A and B is written P(A \cap B) or P(A,B).

Marginal probability or prior probability is the probability of one event, regardless of the other event. Marginal probability is obtained by summing (or integrating, more generally) the joint probability over the unrequired event. This is called marginalization. The marginal probability of A is written P(A), and the marginal probability of B is written P(B).

In these definitions, note that there need not be a causal or temporal relation between A and B. A may precede B or vice versa or they may happen at the same time. A may cause B or vice versa or they may have no causal relation at all. Notice, however, that causal and temporal relations are informal notions, not belonging to the probabilistic framework. They may apply in some examples, depending on the interpretation given to events.

Conditioning of probabilities, i.e. updating them to take account of (possibly new) information, may be achieved through Bayes' theorem.

Contents

[edit] Definition

Given a probability space (Ω,F,P) and two events A, B\in F with P(B) > 0, the conditional probability of A given B is defined by

P(A \mid B) = \frac{P(A \cap B)}{P(B)}.

If P(B) = 0 then P(A \mid B) is undefined.

[edit] Statistical independence

Two random events A and B are statistically independent if and only if

P(A \cap B) \ = \ P(A)  P(B)

Thus, if A and B are independent, then their joint probability can be expressed as a simple product of their individual probabilities.

Equivalently, for two independent events A and B,

P(A|B) \ = \ P(A)

and

P(B|A) \ = \ P(B).

In other words, if A and B are independent, then the conditional probability of A, given B is simply the individual probability of A alone; likewise, the probability of B given A is simply the probability of B alone.

[edit] Mutual exclusivity

Two events A and B are mutually exclusive if and only if A \cap B = \varnothing. Then, P(A \cap B) = 0.

Therefore, if P(B) > 0 then P(A\mid B) is defined and equal to 0.

[edit] Other considerations

  • If B is an event and P(B) > 0, then the function Q defined by Q(A) = P(A | B) for all events A is a probability measure.

[edit] The conditional probability fallacy

The conditional probability fallacy is the assumption that P(A|B) is approximately equal to P(B|A). The mathematician John Allen Paulos discusses this in his book Innumeracy (p 63 et seq), where he points out that it is a mistake often made even by doctors, lawyers, and other highly educated non-statisticians. It can be overcome by describing the data in actual numbers rather than probabilities.

The relation between P(A|B) and P(B|A) is given by Bayes Theorem:

P(B \mid A)= P(A \mid B) \cdot \frac{P(B)}{P(A)}.

[edit] An example

In the following constructed but realistic situation, the difference between P(A|B) and P(B|A) may be surprising, but is at the same time obvious.

In order to identify individuals having a serious disease in an early curable form, one may consider screening a large group of people. While the benefits are obvious, an argument against such screenings is the disturbance caused by false positive screening results: If a person not having the disease is incorrectly found to have it by the initial test, they will most likely be quite distressed until a more careful test shows that they do not have the disease. Even after being told they are well, their lives may be affected negatively.

The magnitude of this problem is best understood in terms of conditional probabilities.

Suppose 1% of the group suffer from the disease, and the rest are well. Choosing an individual at random,

P(disease) = 1% = 0.01 and P(well) = 99% = 0.99.

Suppose that when the screening test is applied to a person not having the disease, there is a 1% chance of getting a false positive result, i.e.

P(positive | well) = 1%, and P(negative | well) = 99%.

Finally, suppose that when the test is applied to a person having the disease, there is a 1% chance of a false negative result, i.e.

P(negative | disease) = 1% and P(positive | disease) = 99%.

Now, calculation shows that:

P(\text{well}\cap\text{negative})=P(\text{well})\times P(\text{negative}|\text{well})=99%\times99%=98.01% is the fraction of the whole group being well and testing negative.
P(\text{disease}\cap\text{positive})=P(\text{disease})\times P(\text{positive}|\text{disease})=1%\times99%=0.99% is the fraction of the whole group being ill and testing positive.
P(\text{well}\cap\text{positive})=P(\text{well})\times P(\text{positive}|\text{well})=99%\times1%=0.99% is the fraction of the whole group having false positive results.
P(\text{disease}\cap\text{negative})=P(\text{disease})\times P(\text{negative}|\text{disease})=1%\times1%=0.01% is the fraction of the whole group having false negative results.

Furthermore,

P(\text{positive})=P(\text{well}\cap\text{positive})+P(\text{disease}\cap\text{positive})=0.99%+0.99%=1.98% is the fraction of the whole group testing positive.
P(\text{disease}|\text{positive})=\frac{P(\text{disease}\cap\text{positive})}{P(\text{positive})}=\frac{0.99%}{1.98%}= 50% is the probability that you actually have the disease if you tested positive.

In this example, it should be easy to relate to the difference between P(positive|disease)=99% and P(disease|positive)= 50%: The first is the conditional probability that you test positive if you have the disease; the second is the conditional probability that you have the disease if you test positive. With the numbers chosen here, the last result is likely to be deemed unacceptable: Half the people testing positive are actually false positives.

[edit] See also

Image:Bvn-small.png Probability distributionsview  talk  edit ]
Univariate Multivariate
Discrete: BenfordBernoullibinomialBoltzmanncategoricalcompound Poissondiscrete phase-typedegenerateGauss-Kuzmingeometrichypergeometriclogarithmicnegative binomialparabolic fractalPoissonRademacherSkellamuniformYule-SimonzetaZipfZipf-Mandelbrot Ewensmultinomialmultivariate Polya
Continuous: BetaBeta primeCauchychi-squareDirac delta functionCoxianErlangexponentialexponential powerFfadingFermi-DiracFisher's zFisher-TippettGammageneralized extreme valuegeneralized hyperbolicgeneralized inverse GaussianHalf-logisticHotelling's T-squarehyperbolic secanthyper-exponentialhypoexponentialinverse chi-square (scaled inverse chi-square) • inverse Gaussianinverse gamma (scaled inverse gamma) • KumaraswamyLandauLaplaceLévyLévy skew alpha-stablelogisticlog-normalMaxwell-BoltzmannMaxwell speedNakagaminormal (Gaussian)normal-gammanormal inverse GaussianParetoPearsonphase-typepolarraised cosineRayleighrelativistic Breit-WignerRiceshifted GompertzStudent's ttriangulartruncated normaltype-1 Gumbeltype-2 GumbeluniformVariance-GammaVoigtvon MisesWeibullWigner semicircleWilks' lambda DirichletGeneralized Dirichlet distribution . inverse-WishartKentmatrix normalmultivariate normalmultivariate Studentvon Mises-FisherWigner quasiWishart
Miscellaneous: bimodalCantorconditionalequilibriumexponential familyInfinite divisibility (probability)location-scale familymarginalmaximum entropyposteriorpriorquasisamplingsingular
Personal tools