Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

Conditional Probability with Discrete Random Variables: Definitions and Formulas, Study notes of Probability and Statistics

Definitions and formulas for conditional probability, a concept used to estimate the probability of an event based on the occurrence of another event. The concept of conditional probability using the 'given' symbol (|) and provides examples and counterexamples. It also discusses the importance of conditional probabilities in various fields such as elections and court cases. The document concludes with some mathematical tools and equations for calculating conditional probabilities.

What you will learn

  • How can we calculate the conditional probability of an event given another event?
  • What is the relationship between conditional probability and independence?
  • What is conditional probability and how is it different from regular probability?

Typology: Study notes

2021/2022

Uploaded on 09/12/2022

karthur
karthur 🇺🇸

4.8

(8)

230 documents

1 / 2

Toggle sidebar

This page cannot be seen from the preview

Don't miss anything!

bg1
By: Neil E. Cotter
PROBABILITY
CONDITIONAL PROBABILITY
Discrete random variables
DEFINITIONS AND FORMULAS
DEF: P(A | B) the (conditional) Probability of A given B occurs
NOT'N: | "given"
EX: The probability that event A occurs may change if we know event B has occurred.
For example, if A it will snow today, and if B it is 90° outside, then knowing that
B has occurred will make the probability of A almost zero. The probability of snow is
higher if we do not know what the temperature is. Thus, P(A | B) < P(A).
DEF: P(A | B) = P(A) A is independent of B the probability of A is unaffected by the
occurrence of event B
EX: Consider two flips of a fair coin. H Heads, and T Tails.
P(H 2nd flip | H 1st flip) = 1/2 = P(H 2nd flip). That is, knowing the outcome of the
first flip doesn't change the probability of the 2nd flip. So the two flips are
independent.
NOTE: Conditional probabilities allow us to improve our estimates of probabilities by
knowing more about the situation we are in. In elections, for example, knowing how
many people are members of each party helps us to improve the accuracy of
predictions about who will win the election. In a court case, knowing more about the
circumstances in which a crime was committed helps us judge the probability of
innocence or guilt.
Conditional probabilities allow us to reduce our sample space to just outcomes in the
event we are conditioning on. For P(A | B), we are finding the probability of A when
the sample space is restricted to B. In a Venn diagram of probabilities, we would
look only inside the area of B, and we would expand the area of B (and everything in
it) to be unity. Our total probabilities of events in B would be unity. P(A | B) would
now correspond to the size of A in B, i.e., AB scaled up by the same factor that
makes the size of B unity.
TOOL: The following formulas define the mathematical behavior of conditional probabilities:
P(A|B)=P(A,B)
P(B)
P(A and B)
P(B)
P(AB)
P(B)
(always true)
P(A|B)=P(A)
(if and only if A and B independent)
P(A,B)=P(A)P(B)
(if and only if A and B independent)
pf2

Partial preview of the text

Download Conditional Probability with Discrete Random Variables: Definitions and Formulas and more Study notes Probability and Statistics in PDF only on Docsity!

By: Neil E. Cotter PROBABILITY CONDITIONAL PROBABILITY Discrete random variables DEFINITIONS AND FORMULAS DEF: P ( A | B ) ≡ the (conditional) Probability of A given B occurs NOT'N: | ≡ "given" EX: The probability that event A occurs may change if we know event B has occurred. For example, if A ≡ it will snow today, and if B ≡ it is 90° outside, then knowing that B has occurred will make the probability of A almost zero. The probability of snow is higher if we do not know what the temperature is. Thus, P ( A | B ) < P ( A ). DEF: P ( A | B ) = P ( A ) ≡ A is independent of B ≡ the probability of A is unaffected by the occurrence of event B EX: Consider two flips of a fair coin. H ≡ Heads, and T ≡ Tails. P ( H 2nd flip | H 1st flip) = 1/2 = P ( H 2nd flip). That is, knowing the outcome of the first flip doesn't change the probability of the 2nd flip. So the two flips are independent. NOTE: Conditional probabilities allow us to improve our estimates of probabilities by knowing more about the situation we are in. In elections, for example, knowing how many people are members of each party helps us to improve the accuracy of predictions about who will win the election. In a court case, knowing more about the circumstances in which a crime was committed helps us judge the probability of innocence or guilt. Conditional probabilities allow us to reduce our sample space to just outcomes in the event we are conditioning on. For P ( A | B ), we are finding the probability of A when the sample space is restricted to B. In a Venn diagram of probabilities, we would look only inside the area of B , and we would expand the area of B (and everything in it) to be unity. Our total probabilities of events in B would be unity. P ( A | B ) would now correspond to the size of A in B , i.e., A∩B scaled up by the same factor that makes the size of B unity. TOOL: The following formulas define the mathematical behavior of conditional probabilities:

P ( A | B ) =

P ( A , B )

P ( B )

P ( A and B ) P ( B )

P ( A ∩ B )

P ( B )

(always true) P ( A | B ) = P ( A ) (if^ and only if^ A^ and^ B^ independent) P ( A , B ) = P ( A ) P ( B ) (if^ and only if^ A^ and^ B^ independent)

By: Neil E. Cotter PROBABILITY CONDITIONAL PROBABILITY Discrete random variables DEFINITIONS, FORMULAS (CONT.) TOOL: Using the Law of Total Probability and the axiom that probabilities of all outcomes in the sample space sum to unity, we can derive additional equations for conditional probability. P ( A ' | B ) = 1 − P ( A | B ) P ( A , B ) = P ( A | B ) P ( B ) = P ( B | A ) P ( A ) P ( A | B ) =

P ( B | A ) P ( A )

P ( B )

(Bayes' theorem) We might be tempted to think that P( A | B ) + P( A | B' ) equals one, but that is not true. Events B and B ' are different universes. Think of one as planet earth, and think of the other as a galaxy far far away. (Our universe consists of just these two places.) The probability of A in the different world may be totally different. What is the relationship between P( A | B ) and P( A | B' )? We put our equations to work. P ( A | B ') =

P ( A , B ')

P ( B ')

P ( A ∩ B ')

1 − P ( B )

P ( A ) − P ( A ∩ B )

1 − P ( B )

which yields an equation involving P ( A | B ) but which requires that we know P ( A ) and P ( B ): P ( A | B ') =

P ( A ) − P ( A , B )

1 − P ( B )

P ( A ) − P ( A | B ) P ( B )

1 − P ( B )

Another interesting form has P ( B | A ) instead of P ( A | B ) but still requires that we know P ( A ) and P ( B ): P ( A | B ') =

P ( A ) − P ( A , B )

1 − P ( B )

P ( A ) − P ( B | A ) P ( A )

1 − P ( B )

= P ( A )

1 − P ( B | A )

1 − P ( B )

Note that this last equation says that if A and B are independent, which means P ( B | A ) = P ( B ), then P ( A | B' ) = P ( A ), which means A and B' are independent.