Skip to content

November 21, 2019

Unit 1 - Introduction to Statistics

Notes from MITx 18.6501x Fundamentals of Statistics covering probability, statistics, LLN, CLT, and fundamental inequalities

By flpvvvv 7 min read Statistics

This is my notes of the course “MITx: 18.6501x Fundamentals of Statistics” on edX. The session that I participated started on Sept 2, 2019. The purpose of this document is to help me remember what I have learnt.

Here is the MITx 18.6501x course page.

Note: a site to calculate/look up statistic curves/values: Keisan Online Calculator

Probability and Statistics

Very explicit explanation of the relation of probability and statistics:

Diagram showing the relationship between probability and statistics

If we know the truth, then we can use “Probability” to predict/explain the “Observations”. However,

statistics is reverse engineering probability.

We have some observations (most of time we call them data), but we have no idea about the truth. We use the observations to estimate the truth. This is the purpose of statistics. From data, we try to recover what the truth is like.

Some conceptions

i.i.d.

Data that i.i.d. stands for independent and identically distributed .

A collection of random variables are i.i.d. if

  1. each follows a distribution , all those distributions are the same, and
  2. are (mutually) independent

Sample average

We denote the sample average, or sample mean, of n random variables by

Laws of Large Numbers (LLN)

Let be i.i.d. random variables, with and .

Laws (weak and strong) of large numbers (LLN):

where the convergence is in probability (as denoted by P on the convergence arrow) and almost surely (as denoted by a.s. on the arrow) for the weak and strong laws respectively.

When n is large enough, the sample average will converge to the expectation of variables.

Central Limit Theorem (CLT)

where the convergence is in distribution, as denoted by (d) on top of the convergence arrow.

Rule of thumb: n 30, to use CTL.

When n is large enough, the sample average converges to a Gaussian distribution.

Three inequalities

Hoeffding’s Inequality

When n is not large enough to apply CTL, we can use Hoeffding’s Inequality.

Given n (n>0) i.i.d. random variables that are almost surely bounded, meaning .

Unlike for the central limit theorem, here the sample size n does not need to be large.

Markov inequality

For a random variable with mean , and any number :

Note that the Markov inequality is restricted to non-negative random variables.

Chebyshev inequality

For a random variable X with (finite) mean and variance , and for any number ,

When Markov inequality is applied to , we obtain Chebyshev’s inequality. Markov inequality is also used in the proof of Hoeffding’s inequality.

Gaussian distribution

It is named after German Mathematician Carl Friedrich Gauss (1777–1855) in the context of the method of least squares (regression).

Gaussian density (PDF)

There is no closed form for their cumulative distribution function (CDF).

Useful properties of Gaussian

Invariant under affine transformation

, then for any ,

Standardization

a.k.a. Normalization/Z-score. If , then

Useful to compute probabilities from CDF of :

Symmetry

if and :

Quantiles

Let in (0,1), the quantile of order of a random variable is the number such that:

Let denote the CDF of :

  • if is invertible, then
  • if , then

Some important quantiles of the are:

2.5%5%10%
1.961.651.28

Three types of convergence

  • is a sequence of random variables
  • T is a random variable (T may be deterministic)

Almost surely (a.s.) convergence

is also known as convergence with probability 1 (w.p.1) and strong convergence.

Convergence in probability

Convergence in distribution

is also known as convergence in law and weak convergence .

for all continuous and bounded function .

When n is large enough, they have the same distribution (the same PDF/CDF).

Properties

  • If converges a.s., then it also converges in probability, and the two limits are equal a.s.

  • If converges in probability, then it also converges in distribution

  • Convergence in distribution implies convergence of probabilities if the limit has a density (e.g. Gaussian):

  • Addition, Multiplication, and Division preserves convergence almost surely (a.s.) and in probability ()

    More precisely, assume

    Then,

    • if in addition, a.s., then

    In general, these rules do not apply to convergence in distribution (d).

Slutsky’s Theorem

For convergence in distribution, the Slutsky’s Theorem will be our main tool.

Let be two sequences of r.v., such that:

where is a r.v. and is a given real number (deterministic limit: ). Then,

  • if in addition, a.s., then

Continuous Mapping Theorem

If is a continuous function: