This page contains mathematical rules we’ll use in this course that may be beyond what is covered in a linear algebra course.
Matrix calculus
Definition of gradient
Let be a vector and be a function of .
Then , the gradient of with respect to is
Gradient of
Let be a vector and be a vector, such that is not a function of .
The gradient of with respect to is
Gradient of
Let be a vector and be a matrix, such that is not a function of .
Then the gradient of with respect to is
If is symmetric, then
Hessian matrix
The Hessian matrix, is a matrix of partial second derivatives
Expected value
Expected value of random variable
The expected value of a random variable is a weighted average, i.e., the mean value of the possible values a random variable can take weighted by the probability of the outcomes.
Let be the probability distribution of . If is continuous then
If is discrete then
Expected value of vector
Let be a vector of random variables.
Then
Expected value of vector
Let be an matrix of constants and a vector of random variables. Then
Expected value of
Let be an matrix of constants, a vector of constants, and a vector of random variables. Then
Expected value of
Let be an matrix of constants and a matrix. Then
Variance
Variance of random variable
The variance of a random variable is a measure of the spread of a distribution about its mean.
Variance of vector
Let be a vector of random variables. Then
This produced the variance-covariance matrix
Variance of
Let be an matrix of constants and a vector of random variables. Then
Probability distributions
Multivariate normal distribution
Let be a vector of random variables, such that follows a multivariate normal distribution with mean and variance . Then the probability density function of is
Linear transformation of normal random variable
Suppose is a multivariate normal random variable with mean and variance . A linear transformation of is also multivariate normal, such that