Probability: Difference between revisions

From David's Wiki
No edit summary
Line 1: Line 1:
Introductory Probability as taught in [https://www.pearson.com/us/higher-education/program/Ross-First-Course-in-Probability-A-9th-Edition/PGM110742.html Sheldon Ross' book]
Calculus-based Probability
 
==Axioms of Probability==
==Axioms of Probability==
* <math>0 \leq P(E) \leq 1</math>
* <math>0 \leq P(E) \leq 1</math>
Line 18: Line 19:
<math>Var(Y) = E(Var(Y|X)) + Var(E(Y | X)</math>
<math>Var(Y) = E(Var(Y|X)) + Var(E(Y | X)</math>
{{hidden | Proof |}}
{{hidden | Proof |}}
==Convergence==
There are 4 types of convergence typically taught in undergraduate courses.<br>
See [https://en.wikipedia.org/wiki/Convergence_of_random_variables Wikipedia Convergence of random variables]
===Almost Surely===
===In Probability===
* Implies Convergence in distribution
===In Distribution===
* Equivalent to convergence in probability if it converges to a degenerate distribution
===In Mean Squared===


==Delta Method==
==Delta Method==
Line 60: Line 71:
Too many. See [https://en.wikipedia.org/wiki/F-distribution the Wikipedia Page].
Too many. See [https://en.wikipedia.org/wiki/F-distribution the Wikipedia Page].
Most important are Chi-sq and T distribution
Most important are Chi-sq and T distribution
==Textbooks==
* Sheldon Ross' A First Course in Probability
* [https://smile.amazon.com/Introduction-Mathematical-Statistics-Robert-Hogg/dp/0321795431?sa-no-redirect=1 Hogg and Craig's Mathematical Statistics]
* Casella and Burger's Statistical Inference

Revision as of 15:40, 7 November 2019

Calculus-based Probability

Axioms of Probability

  • \(\displaystyle 0 \leq P(E) \leq 1\)
  • \(\displaystyle P(S) = 1\) where \(\displaystyle S\) is your sample space
  • For mutually exclusive events \(\displaystyle E_1, E_2, ...\), \(\displaystyle P\left(\bigcup_i^\infty E_i\right) = \sum_i^\infty P(E_i)\)

Monotonicity

  • For all events \(\displaystyle A\), \(\displaystyle B\), \(\displaystyle A \subset B \implies P(A) \leq P(B)\)
Proof

Expectation and Variance

Some definitions and properties.

Total Expection

Dr. Xu refers to this as the smooth property. \(\displaystyle E(X) = E(E(X|Y))\)

Proof

Total Variance

This one is not used as often on tests as total expectation \(\displaystyle Var(Y) = E(Var(Y|X)) + Var(E(Y | X)\)

Proof

Convergence

There are 4 types of convergence typically taught in undergraduate courses.
See Wikipedia Convergence of random variables

Almost Surely

In Probability

  • Implies Convergence in distribution

In Distribution

  • Equivalent to convergence in probability if it converges to a degenerate distribution

In Mean Squared

Delta Method

See Wikipedia

Limit Theorems

Markov's Inequality

Chebyshev's Inequality

Central Limit Theorem

Very very important. Never forget this.
For any distribution, the sample mean converges in distribution to normal. Let \(\displaystyle \mu = E(x)\) and \(\displaystyle \sigma^2 = Var(x)\)
Different ways of saying the same thing:

  • \(\displaystyle \sqrt{n}(\bar{x} - \mu) \sim N(0, \sigma^2)\)
  • \(\displaystyle \frac{\sqrt{n}}{\sigma}(\bar{x} - \mu) \sim N(0, 1)\)
  • \(\displaystyle \bar{x} \sim N(\mu, \sigma^2/n)\)

Law of Large Numbers

Relationships between distributions

This is important for tests.
See Relationships among probability distributions.

Poisson Distributions

Sum of poission is poisson sum of lambda.

Normal Distributions

  • If \(\displaystyle X_1 \sim N(\mu_1, \sigma_1^2)\) and \(\displaystyle X_2 \sim N(\mu_2, \sigma_2^2)\) then \(\displaystyle \lambda_1 X_1 + \lambda_2 X_2 \sim N(\lambda_1 \mu_1 + \lambda_2 X_2, \lambda_1^2 \sigma_1^2 + \lambda_2^2 + \sigma_2^2)\) for any \(\displaystyle \lambda_1, \lambda_2 \in \mathbb{R}\)

Gamma Distributions

Note exponential distributions are also Gamma distrubitions

  • If \(\displaystyle X \sim \Gamma(k, \theta)\) then \(\displaystyle \lambda X \sim \Gamma(k, c\theta)\).
  • If \(\displaystyle X_1 \sim \Gamma(k_1, \theta)\) and \(\displaystyle X_2 \sim \Gamma(k_2, \theta)\) then \(\displaystyle X_2 + X_2 \sim \Gamma(k_1 + k_2, \theta)\).
  • If \(\displaystyle X_1 \sim \Gamma(\alpha, \theta)\) and \(\displaystyle X_2 \sim \Gamma(\beta, \theta)\), then \(\displaystyle \frac{X_1}{X_1 + X_2} \sim B(\alpha, \beta)\).

T-distribution

Ratio of normal and squared-root of Chi-sq distribution yields T-distribution.

Chi-Sq Distribution

The ratio of two normalized Chi-sq is an F-distributions

F Distribution

Too many. See the Wikipedia Page. Most important are Chi-sq and T distribution

Textbooks