When analyzing real-world data—be it financial time series, engineering measurements, public health metrics, or e-commerce behaviors—we rarely focus on a single variable. More often, we observe multiple attributes or features at once: daily returns of several stocks, temperature and humidity at a weather station, or user behaviors across different pages on a website. Random vectors are the probability-theory construct that let us handle these variables jointly, ensuring we capture both individual behaviors and collective interactions among them.
Table of contents
Open Table of contents
1. Introduction: The Need for Random Vectors
1.1 Beyond One-Dimensional Analysis
Traditionally, introductory probability courses start with a single random variable—describing phenomena like one measurement at a time (e.g., an individual’s height, or the outcome of a single die roll). However, real-world problems typically involve multiple measurements:
- Finance: A portfolio might contain multiple assets whose returns we track over time. The daily or monthly return of each asset is a random variable, but we also need to understand how they move together (are they positively correlated, negatively correlated, or nearly independent?).
- Healthcare: A patient’s blood pressure, cholesterol level, and blood glucose are distinct measurements, yet they can co-vary in interesting ways.
- Marketing and E-Commerce: A user’s behavior might be summarized by page views, session duration, items clicked, or items purchased. Analyzing them independently ignores important cross-relationships that could predict future behavior.
In all these scenarios, we gather data across multiple variables, creating an inherently multidimensional space of possible outcomes. Random vectors formalize this concept by treating each “dimension” as a component of a larger probabilistic entity.
1.2 Random Vectors in Formal Terms
A random vector is a function from a sample space into . Each outcome maps to a point . If we focus solely on , that’s just one random variable, but the vector lets us track all variables simultaneously and preserve their interdependencies.
2. Joint Distributions: The Full Multidimensional Story
2.1 Defining the Joint Distribution
To describe how a random vector behaves, we use its joint distribution. In one dimension, we have a single PMF (probability mass function) or PDF (probability density function). In multiple dimensions, we have:
-
Joint CDF:
More generally,
-
Joint PMF/PDF:
- Discrete: .
- Continuous: .
2.2 Marginal Distributions
From the joint distribution, derive marginals by summing/integrating out the other variables. For example:
Marginals show individual behavior, but not co-movement.
2.3 Why Joint Distributions Matter
- Capture correlations (stocks, health indicators).
- Underlie multivariate regression, PCA, Bayesian networks.
- Essential for simulation and generative models.
3. Conditional Distributions
3.1 Formula
Discrete:
Continuous:
3.2 Applications
- Regression models: .
- Classification: .
3.3 Chain Rule
4. Independence
4.1 Definition
Independent if:
4.2 Pairwise vs Mutual Independence
Mutual independence is stronger than pairwise.
4.3 Independence vs Uncorrelatedness
Zero correlation ≠ independence. Independence demands full factorization.
4.4 Applications
- Naive Bayes classifiers.
- Simplified finance risk models.
- Experimental design.
5. Extended Topics
- Conditional Independence: independence given a third variable.
- Correlation & Partial Correlation: controlling for confounders.
- Graphical Models: Bayesian networks, Markov random fields.
- Approximate Independence: Naive Bayes, PCA.
6. Case Studies
- Medical Diagnosis: Blood test scores and disease status.
- Finance: Portfolio correlation structures.
- Quality Control: Multivariate control charts.
7. Pitfalls
- Simpson’s Paradox.
- Spurious correlations.
- Overlooking nonlinear dependencies.
8. Conclusion
Random vectors and their distributions are foundational for multivariate analysis:
- Handle multiple variables in .
- Joint distributions capture relationships.
- Conditional distributions update beliefs with info.
- Independence simplifies but rarely holds perfectly.
By mastering these, you gain a robust toolkit for analyzing real-world multivariate data.
Check out my full article on Medium