Skip to content
Go back

Random Vectors Continued...

Edit page

Modern probability and statistics often focus on scenarios where we aggregate multiple random variables, measure how these variables co-vary, and then draw conclusions about their long-run behavior. In this short post, we’ll introduce three key ideas:

  1. Summation of Random Variables
  2. Covariance and Correlation
  3. Limit Theorems

These concepts lie at the heart of statistical modeling and help explain why averages and aggregated statistics become so powerful in real-world applications.


1. Sums of Random Variables

1.1 Motivation

We frequently encounter sums of random variables in practical contexts—like the total number of defective parts in a factory batch (summing individual defect indicators), or the portfolio return from several stocks (summing daily returns). Understanding the distribution of these sums is crucial for risk analysis, planning, and decision-making.

1.2 Basic Ideas

pZ(z)=xpX(x)pY(zx).p_Z(z) = \sum_{x} p_X(x)\, p_Y(z - x). fZ(z)=fX(x)fY(zx)dx.f_Z(z) = \int_{-\infty}^{\infty} f_X(x)\, f_Y(z - x) \, dx.

1.3 Illustrative Examples


2. Covariance and Correlation

2.1 Covariance

When working with multiple variables, understanding their co-movement is essential. Covariance between X\footnotesize \mathbf{X} and Y\footnotesize \mathbf{Y} is defined as

Cov(X,Y)=E[(XE[X])(YE[Y])]=E[XY]E[X]E[Y].\mathrm{Cov}(X, Y) = E\bigl[(X - E[X]) (Y - E[Y])\bigr] = E[XY] - E[X]\,E[Y].

2.2 Correlation

Covariance is influenced by scale. Correlation standardizes this measure to a range [1,1]\footnotesize \mathbf{[-1,1]}:

ρX,Y=Cov(X,Y)Var(X)Var(Y).\rho_{X,Y} = \frac{\mathrm{Cov}(X,Y)}{\sqrt{\mathrm{Var}(X)\,\mathrm{Var}(Y)}}.

Correlation is widely used in finance (portfolio risk management) and data science (feature analysis).


3. Limit Theorems

3.1 Law of Large Numbers (LLN)

Suppose X1,X2,\footnotesize \mathbf{X_1, X_2, \dots} are i.i.d. with mean μ\footnotesize \mathbf{\mu}. The Law of Large Numbers says that as n\footnotesize \mathbf{n \to \infty},

1ni=1nXi    μ\frac{1}{n}\sum_{i=1}^n X_i \;\to\; \mu

in probability (Weak LLN) or almost surely (Strong LLN). This underpins why sample averages converge to the true mean.

3.2 Central Limit Theorem (CLT)

The Central Limit Theorem states that if X1,X2,\footnotesize \mathbf{X_1, X_2, \dots} are i.i.d. with finite mean μ\footnotesize \mathbf{\mu} and variance σ2\footnotesize \mathbf{\sigma^2}, then

i=1nXinμσn  d  N(0,1),\frac{\sum_{i=1}^n X_i - n\mu}{\sigma\sqrt{n}} \;\xrightarrow{d}\; \mathcal{N}(0,1),

as n\footnotesize \mathbf{n} grows large. This explains why sums or averages of many independent random variables often look normal.

3.3 Practical Significance


Sums of random variables investigate aggregate quantities—vital in risk analysis, manufacturing, and finance. Covariance and correlation quantify how variables move together. Limit theorems assure that sums and averages follow predictable patterns, justifying much of modern statistical inference.

Together, these topics illuminate why and how simple aggregate statistics can capture the essence of complex processes, and how co-variation and long-run behavior form the foundation of scientific, business, and engineering decisions.

Check out my full article on Medium


Edit page
Share this post on:

Next Post
Introduction to Random Vectors