Probability Note
Date: 2024/05/26Last Updated: 2024-05-27T13:37:35.000Z
Categories: Probability
Tags: Probability
Read Time: 16 minutes
0.1 Contents
- 0.2 Random Variables and Probability Distributions
- 0.2.1 Random Variables
- 0.2.2 Cumulative Distribution Function (CDF)
- 0.2.3 Probability Mass Function (PMF) for Discrete Random Variables
- 0.2.4 Probability Density Function (PDF) for Continuous Random Variables
- 0.2.5 Expectation and Variance of Random Variables
- 0.2.6 Moments of Random Variables
- 0.2.7 Standard Deviation of Random Variables
- 0.2.8 Standard Discrete Distributions
- 0.2.9 Standard Continuous Distributions
- 0.3 Discrete Bivariate Distributions
- 0.4 Continuous Bivariate Distributions
- 0.5 Continuous Multivariate Distributions
- 0.6 Independence of Random Variables
- 0.7 Expectation, Covariance and Correlation of Multiple Random Variables
- 0.8 Conditional Distribution and Expectation
- 0.8.1 Conditional Distribution
- 0.8.1.1 Conditional Probability Mass Function (PMF)
- 0.8.1.2 Conditional Commutative Distribution Function (CDF) for Discrete Random Variables
- 0.8.1.3 Conditional Probability Density Function (PDF) for Continuous Random Variables
- 0.8.1.4 Conditional Commutative Distribution Function (CDF) for Continuous Random Variables
- 0.8.2 Conditional Expectation
- 0.8.3 Conditional Variance for Multiple Random Variables
- 0.8.1 Conditional Distribution
- 0.9 Transformations of Random Variables
- 0.10 Generating Functions
- 0.11 Markov and Chebyshev Inequalities
- 0.12 Multivariate Normal Distribution
- 0.12.1 Joint PDF of Multivariate Normal Distribution
- 0.12.2 Joint Moment Generating Function of Multivariate Normal Distribution
- 0.12.3 Moments of Multivariate Normal Distribution
- 0.12.4 Bivariate Normal Distribution
- 0.12.5 Properties of Multivariate Normal Distribution
- 0.12.6 Condition of Independence of Multivariate Normal Distribution
- 0.12.7 Degenerate Multivariate Normal Distribution
- 0.13 Limiting Behaviors of Sums of Random Variables
0.2 Random Variables and Probability Distributions
0.2.1 Random Variables
A random variable is a function that maps the outcomes of a random process to numerical values . If the is discrete, the random variable is called a discrete random variable. If the is continuous, the random variable is called a continuous random variable.
0.2.2 Cumulative Distribution Function (CDF)
The cumulative distribution function (CDF) of a random variable is defined as
0.2.3 Probability Mass Function (PMF) for Discrete Random Variables
The probability mass function (PMF) of a discrete random variable is defined as
0.2.4 Probability Density Function (PDF) for Continuous Random Variables
The probability density function (PDF) of a continuous random variable is defined as
where is the CDF of .
0.2.5 Expectation and Variance of Random Variables
The expectation of a random variable is defined as
where is the PMF of .
where is the PDF of .
The variance of a random variable is defined as
By the definition of variance, the variance is always non-negative.
Alternatively, the variance can be calculated as
0.2.5.1 Properties of Expectation and Variance
where and are constants, and is the covariance between and .
0.2.6 Moments of Random Variables
The n-th moment of a random variable is defined as
0.2.7 Standard Deviation of Random Variables
The standard deviation of a random variable is defined as
0.2.8 Standard Discrete Distributions
0.2.8.1 Bernoulli Distribution
The Bernoulli distribution is a discrete distribution with two possible outcomes: 0 and 1. The PMF of a Bernoulli random variable is defined as
where is the probability of success.
The expectation and variance of a Bernoulli random variable are
0.2.8.2 Binomial Distribution
The Binomial distribution is a discrete distribution that models the number of successes in a fixed number of independent Bernoulli trials. The PMF of a Binomial random variable is defined as
where is the number of trials, is the number of successes, and is the probability of success.
The expectation and variance of a Binomial random variable are
0.2.8.3 Poisson Distribution
The Poisson distribution is a discrete distribution that models the number of events occurring in a fixed interval of time or space. The PMF of a Poisson random variable is defined as
where is the average rate of events.
The expectation and variance of a Poisson random variable are
0.2.8.4 Geometric Distribution
The Geometric distribution is a discrete distribution that models the number of trials needed to achieve the first success in a sequence of independent Bernoulli trials. The PMF of a Geometric random variable is defined as
where is the number of trials needed to achieve the first success, and is the probability of success.
The expectation and variance of a Geometric random variable are
0.2.9 Standard Continuous Distributions
0.2.9.1 Uniform Distribution
The Uniform distribution is a continuous distribution with a constant probability density function (PDF) over a fixed interval. The PDF of a Uniform random variable is defined as
where and are the lower and upper bounds of the interval.
The expectation and variance of a Uniform random variable are
0.2.9.2 Exponential Distribution
The Exponential distribution is a continuous distribution that models the time between events in a Poisson process. The PDF of an Exponential random variable is defined as
where is the rate parameter.
The expectation and variance of an Exponential random variable are
0.2.9.3 Normal Distribution
The Normal distribution is a continuous distribution that is symmetric and bell-shaped. The PDF of a Normal random variable is defined as
where is the mean and is the standard deviation.
The expectation and variance of a Normal random variable are
We usually write to denote that follows a Normal distribution with mean and variance .
0.2.9.4 Gamma Distribution
The Gamma distribution is a continuous distribution that generalizes the Exponential distribution. The PDF of a Gamma random variable is defined as
where is the rate parameter, is the shape parameter, and is the gamma function.
The expectation and variance of a Gamma random variable are
0.2.9.5 Beta Distribution
The Beta distribution is a continuous distribution that is defined on the interval [0, 1]. The PDF of a Beta random variable is defined as
where and are the shape parameters, and is the beta function.
The expectation and variance of a Beta random variable are
0.3 Discrete Bivariate Distributions
A bivariate distribution is a probability distribution that describes the joint behaviour of two random variables.
0.3.1 Joint Probability Mass Function (PMF)
Given two random variables and , the joint probability mass function (PMF) for discrete random variables is defined as
0.3.2 Marginal Probability Mass Function (PMF)
The marginal probability mass function (PMF) of a random variable is defined as
0.3.3 Conditional Probability Mass Function (PMF)
The conditional probability mass function (PMF) of a random variable given is defined as
0.3.4 Expectation and Variance of Bivariate Distributions
The expectation of a bivariate distribution is defined as
The covariance of two random variables and is defined as
Covariance can sometimes be negative, zero, or positive.
And can be calculated as
The correlation coefficient of two random variables and is defined as
We can prove that .
0.3.5 Independent Random Variables
Two random variables and are independent if and only if
for all and .
If and are independent, then
0.3.6 Uncorrelated Random Variables
Two random variables and are uncorrelated if and only if
If and are uncorrelated, then
Note: Uncorrelated random variables are not necessarily independent.
0.4 Continuous Bivariate Distributions
0.4.1 Joint Cumulative Distribution Function (CDF)
Given two random variables and , the joint cumulative distribution function (CDF) for continuous random variables is defined as
0.4.2 Marginal Cumulative Distribution Function (CDF)
The marginal cumulative distribution function (CDF) of a random variable is defined as
0.4.3 Joint Probability Density Function (PDF)
Given two random variables and , if there exists a function such that
for all Lebesgue-measurable sets , then is the joint probability density function (PDF) of and . And and are called jointly continuous random variables.
By the definition of the joint PDF, we have
And
0.4.4 Marginal Probability Density Function (PDF)
The marginal probability density function (PDF) of a random variable is defined as
0.5 Continuous Multivariate Distributions
0.5.1 Joint Cumulative Distribution Function (CDF)
Given random variables , let the vector , the joint cumulative distribution function (CDF) for continuous random variables is defined as
0.5.2 Joint Probability Density Function (PDF)
Given random variables , let the vector , if there exists a function such that
for all Lebesgue-measurable sets , then is the joint probability density function (PDF) of . And are called jointly continuous random variables.
By the definition of the joint PDF, we have
And
0.5.3 Marginal Probability Density Function (PDF)
The marginal probability density function (PDF) of a random variable is defined as
0.6 Independence of Random Variables
Two random variables and are independent if and only if
for all and .
This can be thought as the joint behaviour of and is the product of the marginal behaviour of and .
The definition can also be formulated in terms of the joint PDF:
To show that two random variables are not independent, we only need to find one pair of and such that the equation does not hold.
0.6.1 Functions of Independent Random Variables
Given two independent random variables and , and a function , and a function , the random variables and are also independent.
0.6.2 Mutual Independence of Random Variables
A set of random variables are mutually independent if and only if
Note: Mutual independence implies pairwise independence, however, the converse is not true.
0.6.3 Identically Independent Random Variables (IID)
A set of random variables are identically independent (IID) if and only if
- They are mutually independent.
- They have the same distribution.
0.6.4 Sum of Random Variables
Given two independent random variables and , the sum of and is defined as
Then, the CDF of can be calculated as
The PDF of can be calculated as
This is called the convolution of the PDFs of and .
0.7 Expectation, Covariance and Correlation of Multiple Random Variables
0.7.1 Expectation of Multiple Random Variables
Let the vector be a set of random variables, and for Lebesgue-measurable functions the expectation of is defined as
0.7.2 Properties of Expectation of Multiple Random Variables
- If and are independent, then for any functions and .
0.7.3 Covariance of Multiple Random Variables
The covariance of two random variables and is defined as
The covariance can be calculated as
0.7.3.1 Properties of Covariance of Multiple Random Variables
0.7.3.2 Cauchy-Schwarz Inequality In Terms of Covariance
This equality holds if and only if and are linearly related.
0.7.3.3 Correlation Coefficient of Multiple Random Variables
The correlation coefficient of two random variables and is defined as
By Cauchy-Schwarz Inequality, we have .
0.7.3.4 Moments of Multiple Random Variables
The n-th (raw) moment of a random variable is defined as , and the n-th central moment of a random variable is defined as .
The joint (raw) moment of random variables and is defined as , and the joint central moment of random variables and is defined as .
0.7.3.5 Expectation and Variance of Multiple Random Variables
Given a set of random variables ,
which is a vector of dimensions vector.
The covariance matrix of is defined as
which is a matrix.
0.8 Conditional Distribution and Expectation
0.8.1 Conditional Distribution
0.8.1.1 Conditional Probability Mass Function (PMF)
Given two random variables and , the conditional probability mass function (PMF) of given is defined as
0.8.1.2 Conditional Commutative Distribution Function (CDF) for Discrete Random Variables
Given two random variables and , the conditional commutative distribution function (CDF) of given is defined as
0.8.1.3 Conditional Probability Density Function (PDF) for Continuous Random Variables
Given two random variables and , the conditional probability density function (PDF) of given is defined as
0.8.1.4 Conditional Commutative Distribution Function (CDF) for Continuous Random Variables
Given two random variables and , the conditional commutative distribution function (CDF) of given is defined as
0.8.2 Conditional Expectation
Given two random variables and , the conditional expectation of given is defined as
We can also define function of as
This is a random variable, and we call this the conditional expectation of given .
0.8.2.1 The Law of Iterated Expectations (The Tower Law)
Given two random variables and , the law of iterated expectations states that
Proof:
0.8.2.2 Law of Total Probability
Given a random variable and an event , the law of total probability states that
0.8.2.3 Wald's Equation
Given a random variable and a stopping time which is an integer valued random variable, then
0.8.2.4 Properties of Conditional Expectation
- If , then .
- If and are independent, then .
- For any function and , .
0.8.3 Conditional Variance for Multiple Random Variables
Given two random variables and , the conditional variance of given is defined as
The conditional variance can be calculated as
Note that the conditional variance is a random variable of .
0.8.3.1 Law of Total Variance
Given a random variable and an event , the law of total variance states that
0.9 Transformations of Random Variables
0.9.1 Support of Probability Density Function (PDF)
Given a random variable with a PDF , the support of is the set of values of where .
0.9.2 Monotonic Transformations
Given a random variable with a PDF , and a function , if is a monotonic function, then the CDF of is
Then, the PDF of is
For non-monotonic transformations, we can break the transformation into monotonic parts.
0.9.3 Transformation of Bivariate Random Variables
Given random variables and with a joint PDF , and functions , where is a one-to-one transformation, and let .
We define , which is the Jacobian determinate of as
Then, the joint PDF of is
Note: The Jacobian determinate satisfy: .
0.9.4 Transformation of Multivariate Random Variables
The theorem from Transformation of Bivariate Random Variables can be generalized to multiple random variables.
Given random variables with a joint PDF , and functions , where is a one-to-one transformation, and let .
We define , which is the Jacobian determinate of as
Then, the joint PDF of is
0.10 Generating Functions
0.10.1 The Moment Generating Function (MGF)
Given a random variable , the moment generating function (MGF) of is defined as
The domain of the MGF is the set of such that exists and is finite.
If the domain does not contain an open neighbourhood of , then we say the MGF does not exist.
0.10.1.1 Example: The Moment Generating Function of the Standard Normal Distribution
Given a random variable that follows the standard normal distribution,
By definition,
0.10.1.2 Example: The Moment Generating Function of the Exponential Distribution
Given a random variable that follows the exponential distribution,
By definition,
0.10.1.3 Properties of the Moment Generating Function
The n-th derivative of the MGF at is:
By the previous property, the Maclaurin series of the MGF is:
Also, if have MGF and , then have MGF .
0.10.1.4 Uniqueness of the Moment Generating Function
Given two random variables and with MGF and , if for all in an open neighbourhood of , then and have the same distribution.
0.10.2 Joint Moment Generating Function (JMGF)
The joint moment generating function (JMGF) of random variables is defined as a function from to :
where and .
If the JMGF exists and is finite on a open neighbourhood of , then we say the JMGF exists.
0.10.2.1 Properties of the Joint Moment Generating Function
If the JMGF exists and is finite on a open neighbourhood of , then it uniquely determines the joint distribution of .
The MGF of can be expressed as:
The joint moment of can be expressed as:
0.10.3 Relation Between the Joint Moment Generating Function and Moment Generating Function
Given random variables with MGF , and JMGF , then are mutually independent if and only if
0.10.4 Sums of Independent Random Variables
Given random variables that are independent, and , then the MGF of is
0.10.5 Probability Generating Function (PGF)
Given a random variable that takes non-negative integer values, the probability generating function (PGF) of is defined as
0.10.5.1 Properties of the Probability Generating Function
The PMF of is uniquely determined by .
The n-th factorial moment of is
Random variables are mutually independent if and only if the joint PGF is
The PGF of sum of independent random variables is
0.10.5.2 Relation of PGF and MGF
Given a random variable that takes non-negative integer values, and the PGF and MGF of , then
0.11 Markov and Chebyshev Inequalities
0.11.1 Markov Inequality
Given a non-negative random variable and , then
Proof:
0.11.2 Chebyshev Inequality
Given a random variable with mean and variance , and , then
Proof:
Define , then is a non-negative random variable, and .
Then,
0.12 Multivariate Normal Distribution
We define the higher dimensional normal distribution as an analog of the one dimensional normal distribution.
We say a random vector follows a multivariate normal distribution if it can be expressed as
where and is a vector of means, is a matrix of constants, and is a vector of independent standard normal random variables.
In convention, we write , and we denote the multivariate normal distribution as
0.12.1 Joint PDF of Multivariate Normal Distribution
If we assume that has full rank, We can use multivariate transformation to derive the joint PDF of :
0.12.2 Joint Moment Generating Function of Multivariate Normal Distribution
If we assume that has full rank, the joint moment generating function of is