Basic Concepts in Probability
Last updated
Last updated
Let denotes a random variable, then is a specific value that might assume.
Therefore,
All continuous random variables possess probability density function, PDF.
We can abbreviate the equation as follows, because it is a normal distribution.
and
If they are independent, then
If they are conditioned, then
Theorem of Total Probability states the following.
We can apply Bayes Rule.
In integral form,
It is perfectly fine to to condition any of the rules on arbitrary random variables, e.g. the location of a robot can inferred from multiple sources of random measurements.
Similarly, we can condition the rule for combining probabilities of independent random variables on other variables.
However, conditional independence does not imply absolute independence, that is
The converse is neither true, absolute independence does not imply conditional independence.
The expected value of a random variable is given by
Expectation is a linear function of a random variable, we have the following property.
Covariance measures the squared expected deviation from the mean. Therefore, square root of covariance is in fact variance, i.e. the expected deviation from the mean.
Theorem of Total Probability states the following.
We can apply Bayes Rule.
In integral form,
It is perfectly fine to to condition any of the rules on arbitrary random variables, e.g. the location of a robot can inferred from multiple sources of random measurements.
Similarly, we can condition the rule for combining probabilities of independent random variables on other variables.
However, conditional independence does not imply absolute independence, that is
The converse is neither true, absolute independence does not imply conditional independence.
The expected value of a random variable is given by
Expectation is a linear function of a random variable, we have the following property.
Covariance measures the squared expected deviation from the mean. Therefore, square root of covariance is in fact variance, i.e. the expected deviation from the mean.
However, in general, is not a scalar value, it is generally a vector. Let be a positive semi-definite and symmetric matrix, which is a covariance matrix.
The joint distribution of two random variables and can be described as follows.
If , then
If is a quantity that we would like to infer from , the probability is referred as prior probability distribution and is called data, e.g. laser measurements. is called posterior probability distribution over .
In robotics, is called generative model. Since does not depend on , is often written as a normalizer in Bayes rule variables.
for as long as .
Finally, entropy of a probability distribution is given by the following expression. Entropy is the expected information that the value of carries.
In the discrete case, the is the number of bits required to encode x using an optimal encoding, assuming that is the probability of observing .
If is a quantity that we would like to infer from , the probability is referred as prior probability distribution and is called data, e.g. laser measurements. is called posterior probability distribution over .
In robotics, is called generative model. Since does not depend on , is often written as a normalizer in Bayes rule variables.
for as long as .
Finally, entropy of a probability distribution is given by the following expression. Entropy is the expected information that the value of carries.
In the discrete case, the is the number of bits required to encode x using an optimal encoding, assuming that is the probability of observing .