Describe and distinguish between continuous and discrete random variables.
Define and distinguish between the probability density function, the cumulative distribution function, and the inverse cumulative distribution function.
Calculate the probability of an event given a discrete probability function.
Distinguish between independent and mutually exclusive events.
Define joint probability, describe a probability matrix, and calculate joint probabilities using probability matrices.
Define and calculate a conditional probability and distinguish between conditional and unconditional probabilities.
Interpret and apply the mean, standard deviation, and variance of a random variable.
Calculate the mean, standard deviation, and variance of a discrete random variable.
Interpret and calculate the expected value of a discrete random variable.
The expected value E[X] is very closely related to the mean value of a variable as well as the fair price of a payoff distribution. While the mean is more focused on backward-looking descriptions, the expected value is more forward-looking with stronger assumptions around the data generation process (data was and will be generated by same process). The expected value operator is linear, thus: E[X+Y]=E[X]+E[Y] and E[cX]=cE[X]. However, it is not multiplicative: E[X^2 ]≠E[X]^2.
Calculate and interpret the covariance and correlation between two random variables.
Calculate the mean and variance of sums of variables.
Describe the four central moments of a statistical variable or distribution: mean, variance, skewness, and kurtosis.
The concept of central moments can be generalised as follows: μ_k=E[(X-μ)^k]. The central moment can be standardised by dividing it with σ^k.
Interpret the skewness and kurtosis of a statistical distribution, and interpret the concepts of coskewness and cokurtosis.
Describe and interpret the best linear unbiased estimator.
The best linear unbiased estimator (BLUE) is the estimator with the lowest variance for an unknown population metric (e.g. mean) that can be represented in a linear function and is unbiased. Unbiased refers to the estimator’s expected value being equal to the true value of the unknown metric.
Uniform distribution
The uniform distribution resembles a constant probability density function between a defined lower (b1) and upper (b2) bound. For the bounds being 0 and 1, respectively, the distribution is called standard uniform distribution.
Bernoulli distribution
The Bernoulli distribution shows a single realisation of a Bernoulli variable, which is 1 with probability p and 0 with probability q = 1 – p. The distribution will show up when modelling occurrences of binary events, such as bond defaults, whether a return will be positive or negative, or whether rates fall or rise.
Binomial distribution
The Binomial distribution can be considered as the collection of realisations of multiple Bernoulli variables, for example the defaults of different bonds. For k variables, the distribution will yield the probability of n out of k variables realising a value of 1.
Poisson distribution
The Poisson distribution is often used to model the occurrence of events over time, e.g. the number of bond defaults in a portfolio. It can be used to model jumps in jump-diffusion models.
Normal distribution
The normal distribution (also called Gaussian distribution) has a symmetrical probability density function, with the mean and median coinciding at the maximum of this PDF. Formally, the normal distribution is depicted by X~N(μ,σ^2). Normally distributed log returns are used across finance, for example is the Black-Scholes option pricing model.
Lognormal distribution
The lognormal distribution allows us to model standard returns directly, without applying a log to them. This applies because the log of a lognormally distributed variable has a normal distribution. The lognormal distribution is especially useful for modelling returns (1+R) because is limits the left side to R=-100%, which is applicable to all limited liability financial instruments.
Chi-squared distribution
The Chi-squared distribution is achieved by combining the square of k standard normal variables. It can only depict positive values (because of the square) and is thus asymmetrical. The Chi-squared distribution is used especially for hypothesis testing.
Student’s-t distribution
Student’s t-distribution is widely used for modelling returns of financial assets and in hypothesis testing. With increasing k, the t-distribution will converge towards the standard normal distribution. The t-distribution is symmetrical around its mean, which is 0. The variance is given by k/(k-2) for k > 2, and converges toward 1 with increasing k.
F-distribution
The F-distribution is based on two independent chi-squared distributions U1 and U2 with k1 and k2 degrees of freedom. With increasing k1 and k2, the mean and mode of the distribution will converge towards 1, with the F-distribution converging towards the normal distribution. The square of a variable X with a t-distribution will have an F-distribution. If X is a random variable with a t-distribution and k degrees of freedom, then X2 has an F-distribution with 1 and k degrees of freedom: X^2~F(1,k).
Describe the central limit theorem and the implications it has when combining independent and identically distributed (i.i.d.) random variables.