Probability

Axioms

  • ES.p(E)0
  • p(S)=1
  • p(AB)=p(A)+p(B) if A and B are disjoint events
  • p(Ec)=1p(E)
  • Objective probabilities: p(A)=limNNAN observation
  • Frequentist Bayesian statistics
  • Subjective probabilities: theoretical estimates for the probabilities using a model
  • Computing probabilities: S discrete and finite, S={x1,x2,,xN}, assume p({xi})=1N, 1<i<N. Then, p(E)=#outcomesinE#outcomesinS.

Combinatorics

Exercises

Question 1

How many distinct ways you can arrange the 24 letters of the alphabet? W=24!

N distinct objects can be arranged in W=N! ways.

Letters in ’WHAT’: W=4!

Question 2

Letters in ’CHEESE’: W=6!/3! = Number of ways to arrange it/number of ways to arrange the non-distinct letters

Question 3

Letters in ’FREEZER’: W=7!/(3!2!1!1!).

3! from the Es, 2! from the Rs, 1! from the F, 1! from the Z.

Multinomial coefficient gives the number of arrangements W of N objectects from k distinct categories each of which appears Nj times, where N=j=1kNj. W=N!N1!Nk!=N!jNj!.

We get a special case for k=2, the binomial coefficient. W=N!N1!N2!=N!N1!(NN1)!=N choose N1. (W=NM)

Question 4

For N coin tosses, how many are tails?

NA:= Outcome A in N trials.

We want pN(NT).

For event E: NT=5, N=12. Then, WE is 12 choose 5. WS=(2!)12. P(5)=(125)212.

So, P(NT)=WEWS=N!NT!(NNT)!(2!)N=1(pT)NT(1pT)NNT((NNT)).

Binomial Distribution

2 outcomes with pa,pb=1pa in N trials, pN(Na)=(pa)Na(1pa)NNa(NNa).

Multinomial Distribution

N trials with k outcomes with probabilities p1,,pk. The probability of finding N1,N2,,Nk outcomes j with N=j=1kNj. pN(N1,N2,,Nk)=N!j=1k1Nj!pjNj.

Stirling’s Approximation

lnN!=NlnNN

N!=exp(NlnNN)=NNexp(N)=(Ne)N.

Stirling’s: N!=2πN(Ne)N.

How? N!=xnexp(x)dx=Γ(n+1).

(N+1/2)ln(N+1/2)(N+1/2)(1/2ln1/21/2)(NlnNN) (N+1/2)(lnN+ln(1+1/(2N)))N1/21/2ln1/2+1/2NlnN+N 1/2lnN+(N+1/2)ln(1+1/(2N))1/2ln1/2 lnN+ 01/2ln1/2 Thus, the correction is of the order N, which is what Stirling’s formula has.

Ways

W=N!j=1kNj!, j=1kNj=N. pj=Nj/N.

Approximating

Ways

W≈=2πN(Ne)N1iNi!=1jpjNj

lnW=jNjlnpj

lnWN=jpjlnpj=SNkBS=kBlnW=kBjpjlnpj. Thus, we recovered our typical entropy! This makes sense since Entropy is related to the number of ways a system can be arranged.

The entropy of a fair dice is: S=kBln6.

No knowledge implies fiar dice maximizes the entropy.

Lets say we have a maximally unfair dice p6=1,pi6=0. Then the entropy is minimized. Assume we have no missing information.

Lets say someone says, Ndice=3.5. Then we would assume it is a fair dice.

If someone says Ndice=3.0 we would assume it is an unfair dice (weighted lower). There is a distribution of probabilities for the individual dice probabilities to be:

  • p3=1, pi3=0
  • p2=p4=12, pi2,4=0

Thus, there is an increase in missing information.

Author: Christian Cunningham

Created: 2024-05-30 Thu 21:20

Validate