« Back to Statistics

APPROXIMATING DISTRIBUTIONS


Contents:


0. Key ideas and definitions

Some phrases and terms regarding probability distributions are discussed in the file on quantifying probabilities. Chiefly, the phrase “drawing from a distribution” is clarified there. Other key ideas and concepts are…

0.1. Product measure

$M_1 \bigotimes M_2$ denotes the product measure of $M_1$ and $M_2$, given that $M_1$ and $M_2$ are measures defined on sets $X$ and $Y$ respectively. $M_1 \bigotimes M_2$ is the product measure defined on $X \times Y$ such that:

$(M_1 \bigotimes M_2)(A \subseteq X \times Y) = M_1(A)M_2(A)$

NOTE: We denote:

0.1.1. Product measure of a probability measure

A product measure of a probability measure is the product measure of two or more copies of the probability measure, and results in an independent joint probability distribution (independent means that no component of the joint outcome affects the others); this represents the joint distribution of tuples of $n$ independently drawn samples.

0.2.1. Support of a distribution

The set of values for which the mass (for discrete distributions) or density (for continuous distributions) is non-zero. Furthermore, we can define the support of a probability distribution as the smallest set of values for which the probability mass is $1$.

0.2.2. Spread of a distribution

The deviation of distributed values from the distribution mean. This can be measured in many ways, such as variance, standard deviation, mean deviation, interquartile range, etc. In a probability distribution, convergence of the spread of the distribution to a specific interval refers to the asymptotic rise with respect to some variable $k$ in the proportion of the mass of the distribution concentrated in the interval. Asymptotic rise in proportion means that the proportion capable of reaching arbitrarily close to $1$, given a large enough $k$.

0.2.3. Sum of two distributions

Given:

We define:

$\mathbb{P} + … + \mathbb{P}$ ($n$ times) $= +_*(\mathbb{P}^n)$

NOTE: We denote:

0.2.4. Distribution parameter functions

$\mu$:

Distribution mean function, i.e. a function that inputs a probability measure defined on a given set $S$ and outputs the corresponding distribution’s mean (if it exists).

$Var$:

Distribution variance function, i.e. a function that inputs a probability measure defined on a given set $S$ and outouts the corresponding distribution’s variance (if it exists).

0.3 Maintaing probability distribution across samples

Henceforth, “distribution” = “probability distribution” unless specified.

Maintaining the distribution of a random process across samples
The probability distribution of a random process is maintained across samples, i.e. for each sample only if each outcome of the process is independent of past outcomes. Maintaining the probability distribution for each sample is needed when studying this distribution through its samples.

Maintaining the distribution of a random selection across selections
Note that a random selection is a kind of random process. Any sequence of random selections from a population are independent if they are made with replacement (more on sampling with replacement later), i.e. if at each selection, each set of metric values have a probability of being observed that stays the same regardless of whether they were observed before. Note that practically, for a large enough population with respect to the number of samples taken, the distribution from random sampling without replacement can still be generalised as a distribution from a random process whose distribution is maintained across samples, given the negligible (near-zero) change in the distribution of the metric being studied among the population’s individuals even as samples are drawn without replacement.

1. Estimators

Henceforth, “distribution” = “probability distribution” unless specified.

An estimator is a collection of maps (collection, because there is a map for each value of $n \in \mathbb{N}$) $T^n:X^n \rightarrow \mathbb{R}$ which implements some operation on any tuple of $X^n$. The purpose of an estimator is to estimate some real value related to the probability measure on $X$, such as the distribution mean, variance, etc.

NOTE: Independent and identically distributed (IID) samples as a prerequisite for estimation:
The essence of estimation is averaging; we average over many samples from a distribution to get an idea about the distribution itself. To do this, we must first ensure a lack of dependence between consequent samples, since such dependence changes the distribution of consequent samples. In the same vein, since estimation depends on averaging samples from the same distribution, we need to ensure the samples are drawn from the same distribution, i.e. we need to ensure the samples are identically distributed.

1.1. Distribution of an estimator

Consider a theoretical distribution $\mathbb{P}$ which models a given random process whose sample space is $X$. Given an estimator $T^n:X^n \rightarrow \mathbb{R}$ (for all $n$) which estimates some parameter of $\mathbb{P}$, the distribution the estimator is essentially the pushforward measure of $\mathbb{P}^n$ through $T^n$, i.e. $T^n_*\mathbb{P}^n$. This can be understood as the distribution of the results of applying a function $T^n$ to $n$ samples each drawn from $\mathbb{P}$.

1.2. Properties of estimators

These properties support the purpose of estimators; an estimator without one or both of these properties cannot be used for an accurate estimation of distribution parameters. Now, given we have an estimator $T^n$ meant to estimate the parameter $\phi(\mathbb{P})$ of the distribution $\mathbb{P}$, we want the following properties for $T^n$:

1. Unbiasedness:

$\mu({T^n}_*\mathbb{P}^n) = \phi(\mathbb{P})$

2. Consistency:

$\displaystyle \lim_{n \rightarrow \infty} T^n_*\mathbb{P}^n([\phi(\mathbb{P}) - \epsilon, \phi(\mathbb{P}) + \epsilon]) = 1, \forall \epsilon > 0$

In words, unbiasedness implies that the mean of the distribution of the estimator values equals to the parameter to be estimated. Consistency implies that as the number of samples taken rises, the mass of the distribution of estimator values converges to an arbitrarily small neighbourhood around the parameter to be estimated.

1.3. Some noteworthy estimators

  1. Sample mean, distributed by $\bar{\mathbb{P}}_n$
  2. Sample variance, distributed by $\frac{1}{n}(\mathbb{P}-\mu(\mathbb{P})^n)$

MORE ON SAMPLE MEAN DISTRIBUTION:
The sample mean distribution is such that….

2. Limit laws

Limit laws are results about the convergence of the sample mean distribution as the number of samples taken increases. Why does this matter? The distribution over which we take sample means may be any distribution, including the distribution of an estimator values. Since the essence of estimation is averaging, it helps to know how averages behave as the number of samples taken increases. Note that, as with estimators, all limit laws presuppose that (1) samples are drawn independently and (2) samples are drawn from the same distribution. In other words, we only deal with independently and identically distributed (IID) samples.

2.1. Laws of large numbers

These laws are results about the convergence of spread of the sample mean distribution.

2.1.1. Weak law of large numbers

$\displaystyle \lim_{n \rightarrow \infty} \bar{\mathbb{P}}_n([\mu(\mathbb{P}) - \epsilon, \mu(\mathbb{P}) + \epsilon]) = 1, \forall \epsilon > 0$

In words, as the number of samples taken $n$ taken from a given distribution increases, the spread of the sample mean distributions converges to an arbitrarily small interval around the given distribution’s mean. In other words, averaging over larger numbers of samples leads to more precise estimates.

2.1.2. Strong law of large numbers

$\displaystyle \mathbb{P}^{\infty}({(x_1, x_2 …) : \lim_{n \rightarrow \infty}\frac{1}{n}\sum_{i=1}^n x_i = \mu(\mathbb{P})}) = 1$

NOTE: $\mathbb{P}^{\infty} = \mathbb{P} \bigotimes \mathbb{P} \bigotimes …$

In words, given infinite (i.e. potentially limitless) IID samples from a given distribution, it is certain that sample means of any number of these samples taken converges to the given distribution’s mean as the number of these samples taken rises. Another way to word the law is to say that the set of infinite (i.e. potentially limitless) IID samples from a given distribution is exactly the set of all values whose subsets’ sample means converge to the given distribution’s mean as the subsets’ size rises. This implies that sample mean of any sequence of IID samples will certainly converge eventually to the given distribution’s mean as more IID samples are added to the sequence.

2.2. Central limit theorem

This is a result about the convergence of the sample mean distribution as a whole to a normal distribution, as the number of samples taken rises. In essence, this result helps approximate the sample mean distribution as a normal distribution, especially when the number of samples taken is very large.

$\displaystyle \lim_{n \rightarrow \infty} \sqrt{n}(\bar{\mathbb{P}}_n - \mu(\mathbb{P}))((-\infty, t]) = \text{Normal}(0, \sqrt{Var(\mathbb{P})})((-\infty, t])$


We can reformulate the above in less precise terms as:

NOTE: “ $\approx$ “ here means “approximately the same as”.

The purpose of the above reformulations is to help understand what the central limit theorem means and how it could be used in approximating a sample mean distribution. The last three formulations may be less precise, but they show that it is the sample mean distribution that is being estimated, and that can be approximated directly if needed.