1. Logic Probability and Uncertainty

Deductive Logic and Plausible Reasoning

A is trueB is true

If we know A is true then when can deduce that B is true. If B is not true we can deduce A is not true.

Pasted image 20250506122112.webp|292

A and B both true is the region AB. A or B are true is the region AB.

If A is not true we cannot deduce anything about B since BA is nonempty.

The same thing with B is true, we can't deduce anything about A though it increases the plausibility that A is true. When changing the plausibility of one proposition based on another proposition we are making an induction.

Desired Properties of Plausibility Measures:

  1. Degrees of plausibility are nonnegative
  2. They qualitatively agree with common sense. Larger means greater plausibility
  3. If a proposition can be represented in multiple ways, then they must have the same plausibility
  4. Always take all relevant evidence into account
  5. Equivalent states of knowledge are always given the same plausibility

Probability

Random Experiment - Outcome is not completely predictable. Can be repeated under the same conditions.
Outcome - the result of one single trial of the single trial of the random experiment
Sample Space - set of all possible outcomes of a single trial denoted Ω or U for universe.
Event - any set of possible outcomes of a random experiment
union - either event A or event B
intersection - both event A and event B
complement - set of outcomes not in A denoted here as A~
mutually exclusive - AB= (disjoint)
partition - Bi=Ω and BiBj= , ij

Axioms of Probability

  1. P(A)0 for any event A
  2. P(Ω)=1
  3. If A and B are mutually exclusive then P(AB)=P(A)+P(B)

basic probability theorems

Basic Probability Theorems

P(Ω)=B which is the powerset of Ω

1.3.1

AB,P(A)=1P(Ac)
pf:
Ω=AAc,AAc=
1=P(Ω)=P(AAc)=P(A)+P(Ac)

1.3.2

P()=0
pf:
P()=1P(c)=1P(Ω)

1.3.3

ABP(A)P(B)
pf:
B=A(BAc) and A(BAc)=

P(B)=P(A(BAc))=P(A)+P(BAc)

1.3.4

AB,0P(A)1

pf:
AΩP(A)P(Ω) and P()P(A)

1.3.5 Inclusion Exclusion

A,BΩP(AB)=P(A)+P(B)P(AB)
pf:
P(B)=P(AB)+P(AcB)
AB=A(AcB)P(AB)=P(A)+P(AcB)=P(A)+P(B)P(AB)

Continuity Theorem

limnP(An)=P(limnAn)

1.3.6

Let {An} be non-decreasing
limn(An)=P(limnAn)=P(n=1An)
Let {An} be non-increasing
limnP(An)=P(limnAn)=P(n=1An)
pf (non-decreasing):
R1=A1R2=A2A1Rn=AnAn1P(limnAn)=P(n=1An)=P(n=1Rn)=n=1P(Rn)=limkn=1kP(Rn)=limk{P(A1)+n=2k[P(An)P(An1)]}=limnP(An)

1.3.7 Boole's Inequality

Let {An} be an arbitrary sequence of events

P(n=1An)n=1P(An)
pf:

{Dn=i=1nAiDn is non-decreasingDj=Dj1AjP(Dj)P(Dj1)+P(Aj)P(Dj)P(Dj1)P(Aj)P(n=1An)=P(n=1Dn)=limn{P(D1)+j=1n[P(Dj)P(Dj1)]}limn{P(A1)+j=2nP(Aj)}=limnj=1nP(Aj)=n=1P(An)

Marginal Probability - P(A)=P(AB)+P(ABc)

Joint Probability and Independence

Conditional probability of event B relative to the hypothesis of event A

P(A|A)=1 and P(B|A)=P(AB|A)P(AB|A)P(A|A)=P(AB)P(A)

Conditional Probability

Let A and B be events where P(A)>0

Pasted image 20250516160854.webp|230
P(B|A)=P(AB)P(A)where:1. P(B|A)02. P(A|A)=13. P(n=1Bn|A)=n=1P(Bn|A)B1,B2, are mutually exclusive

Multiplication Rules

P(AB)=P(B|A)P(A)P(AB)=P(BA)=P(A|B)P(B)P(ABC)=P((AB)C)=P(C|AB)P(AB)=P(C|AB)P(B|A)P(A)

Total Law of Probability

P(B)=i=1kP(Ai)P(B|Ai)
where Ai,i=1,2,3,,k forms a partition of Ω

pf.
P[BΩ]=P[B(A1A2Ak)]=P[(BA1)(BAk)]=P[(B|A1)]+P[B|A2]++P[B|Ak]

Bayes Theorem

Assume A1,A2,,Ak be events such that P(Ai)>0 and they form a partition.
For any event B:
P(Aj|B)=P(Aj)P(B|Aj)i=1kP(Ai)P(B|Ai)
pf:
P(Aj|B)=P(BAj)P(B)=P(AjB)P(B)=P(Aj)P(B|Aj)P(B)

Independence

A and B are independent if:
P(AB)=P(A)P(B)
Equivalently if:
P(B|A)=P(B)
Mutually Independent iff they are pairwise independent:

P(AiAj)=P(Ai)P(Aj)and for any collectionP(Ad1Ad2Adk)=P(Ad1)P(Ad2)P(Adk)

Baye's Theorem

Let B1,B2,,Bn be a set of unobservable events which partition the universe.

We start with our prior probability P(Bi) for each which can be informative of non-informative. This distribution gives the weighting of our prior beliefs. Then we find that A has occurred.

The Likelihood of the unobservable events B1,,Bn is the conditional probability that A has occurred given each Bi, P(A|Bi). It's the weight given to each of the Bi events given that A occurred.

The posterior probability is P(Bi|A), this distribution contains the weight we attach to each of the evens Bi after we know event A has occurred.

Bayesian Universe

Has two dimensions, one observable and one unobservable where the observable dimension is the horizontal.

Pasted image 20250516171305.webp|219

Pasted image 20250516171330.webp|220

Odds

odds(A)=P(A)P(Ac)

Bayes Factor

B×prior odds=posterior oddsB=P(D|C)P(D|Cc)

D= data that occurred, C = outcome in question