From Surf Wiki (app.surf) — the open knowledge base
Sum of normally distributed random variables
Aspect of probability theory
Aspect of probability theory
In probability theory, calculation of the sum of normally distributed random variables is an instance of the arithmetic of random variables.
This is not to be confused with the sum of normal distributions which forms a mixture distribution.
Independent random variables
Let X and Y be independent random variables that are normally distributed (and therefore also jointly so), then their sum is also normally distributed. i.e., if
:X \sim N(\mu_X, \sigma_X^2) :Y \sim N(\mu_Y, \sigma_Y^2) :Z=X+Y,
then
:Z \sim N(\mu_X + \mu_Y, \sigma_X^2 + \sigma_Y^2).
This means that the sum of two independent normally distributed random variables is normal, with its mean being the sum of the two means, and its variance being the sum of the two variances (i.e., the square of the standard deviation is the sum of the squares of the standard deviations).{{Citation
In order for this result to hold, the assumption that X and Y are independent cannot be dropped, although it can be weakened to the assumption that X and Y are jointly, rather than separately, normally distributed. (See here for an example.)
The result about the mean holds in all cases, while the result for the variance requires uncorrelatedness, but not independence.
Proofs
Proof using characteristic functions
The characteristic function
:\varphi_{X+Y}(t) = \operatorname{E}\left(e^{it(X+Y)}\right)
of the sum of two independent random variables X and Y is just the product of the two separate characteristic functions:
:\varphi_X (t) = \operatorname{E}\left(e^{itX}\right), \qquad \varphi_Y(t) = \operatorname{E}\left(e^{itY}\right)
of X and Y.
The characteristic function of the normal distribution with expected value μ and variance σ2 is
:\varphi(t) = \exp\left(it\mu - {\sigma^2 t^2 \over 2}\right).
So
: \begin{align} \varphi_{X+Y}(t)=\varphi_X(t) \varphi_Y(t) & =\exp\left(it\mu_X - {\sigma_X^2 t^2 \over 2}\right) \exp\left(it\mu_Y - {\sigma_Y^2 t^2 \over 2}\right) \[6pt] & = \exp \left( it (\mu_X +\mu_Y) - {(\sigma_X^2 + \sigma_Y^2) t^2 \over 2}\right). \end{align}
This is the characteristic function of the normal distribution with expected value \mu_X + \mu_Y and variance \sigma_X^2+\sigma_Y^2
Finally, recall that no two distinct distributions can both have the same characteristic function, so the distribution of X + Y must be just this normal distribution.
Proof using convolutions
For independent random variables X and Y, the distribution f**Z of Z = X + Y equals the convolution of f**X and f**Y:
:f_Z(z) = \int_{-\infty}^\infty f_Y(z-x) f_X(x) , dx
Given that f**X and f**Y are normal densities,
: \begin{align} f_X(x) = \mathcal{N}(x; \mu_X, \sigma_X^2) = \frac{1}{\sqrt{2\pi}\sigma_X} e^{-(x-\mu_X)^2/(2\sigma_X^2)} \[5pt] f_Y(y) = \mathcal{N}(y; \mu_Y, \sigma_Y^2) = \frac{1}{\sqrt{2\pi}\sigma_Y} e^{-(y-\mu_Y)^2/(2\sigma_Y^2)} \end{align}
Substituting into the convolution:
: \begin{align} f_Z(z) &= \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}\sigma_Y} \exp \left[-{(z-x-\mu_Y)^2 \over 2\sigma_Y^2}\right] \frac{1}{\sqrt{2\pi}\sigma_X} \exp \left[-{(x-\mu_X)^2 \over 2\sigma_X^2}\right] , dx \[6pt] &= \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}\sqrt{2\pi}\sigma_X\sigma_Y} \exp \left[-\frac{\sigma_X^2(z-x-\mu_Y)^2 + \sigma_Y^2(x-\mu_X)^2}{2\sigma_X^2\sigma_Y^2}\right] , dx \[6pt] &= \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}\sqrt{2\pi}\sigma_X\sigma_Y} \exp \left[ -\frac {\sigma_X^2(z^2 + x^2 + \mu_Y^2 - 2xz - 2z\mu_Y + 2x\mu_Y) + \sigma_Y^2(x^2 + \mu_X^2 - 2x\mu_X)} {2\sigma_Y^2\sigma_X^2} \right] , dx \[6pt] &= \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}\sqrt{2\pi}\sigma_X\sigma_Y} \exp \left[ -\frac { x^2(\sigma_X^2 + \sigma_Y^2) - 2x(\sigma_X^2(z - \mu_Y) + \sigma_Y^2\mu_X) + \sigma_X^2(z^2 + \mu_Y^2 - 2z\mu_Y) + \sigma_Y^2\mu_X^2 } {2\sigma_Y^2\sigma_X^2} \right] , dx \[6pt] \end{align}
Defining \sigma_Z = \sqrt{\sigma_X^2 + \sigma_Y^2}, and completing the square:
: \begin{align} f_Z(z) &= \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}\sigma_Z} \frac{1}{\sqrt{2\pi}\frac{\sigma_X\sigma_Y}{\sigma_Z}} \exp \left[ -\frac { x^2 - 2x\frac{\sigma_X^2(z - \mu_Y) + \sigma_Y^2\mu_X}{\sigma_Z^2} + \frac{\sigma_X^2(z^2 + \mu_Y^2 - 2z\mu_Y) + \sigma_Y^2\mu_X^2}{\sigma_Z^2} } {2\left(\frac{\sigma_X\sigma_Y}{\sigma_Z}\right)^2} \right] , dx \[6pt] &= \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}\sigma_Z} \frac{1}{\sqrt{2\pi}\frac{\sigma_X\sigma_Y}{\sigma_Z}} \exp \left[ -\frac { \left(x - \frac{\sigma_X^2(z - \mu_Y) + \sigma_Y^2\mu_X}{\sigma_Z^2}\right)^2 - \left(\frac{\sigma_X^2(z - \mu_Y) + \sigma_Y^2\mu_X}{\sigma_Z^2}\right)^2 + \frac{\sigma_X^2(z - \mu_Y)^2 + \sigma_Y^2\mu_X^2}{\sigma_Z^2} } {2\left(\frac{\sigma_X\sigma_Y}{\sigma_Z}\right)^2} \right] , dx \[6pt] &= \int_{-\infty}^\infty \frac{1}{\sqrt{2\pi}\sigma_Z} \exp \left[ -\frac { \sigma_Z^2\left(\sigma_X^2(z - \mu_Y)^2 + \sigma_Y^2\mu_X^2\right) - \left(\sigma_X^2(z - \mu_Y) + \sigma_Y^2\mu_X\right)^2 } {2\sigma_Z^2\left(\sigma_X\sigma_Y\right)^2} \right] \frac{1}{\sqrt{2\pi}\frac{\sigma_X\sigma_Y}{\sigma_Z}} \exp \left[ -\frac { \left(x - \frac{\sigma_X^2(z - \mu_Y) + \sigma_Y^2\mu_X}{\sigma_Z^2}\right)^2 } {2\left(\frac{\sigma_X\sigma_Y}{\sigma_Z}\right)^2} \right] , dx \[6pt] &= \frac{1}{\sqrt{2\pi}\sigma_Z} \exp \left[ - { (z-(\mu_X+\mu_Y))^2 \over 2\sigma_Z^2 } \right] \int_{-\infty}^{\infty} \frac{1}{\sqrt{2\pi}\frac{\sigma_X\sigma_Y}{\sigma_Z}} \exp \left[ - \frac{\left(x-\frac{\sigma_X^2(z-\mu_Y)+\sigma_Y^2\mu_X}{\sigma_Z^2}\right)^2}{2\left(\frac{\sigma_X\sigma_Y}{\sigma_Z}\right)^2} \right] , dx \end{align}
The expression in the integral is a normal density distribution on x, and so the integral evaluates to 1. The desired result follows:
:f_Z(z) = \frac{1}{\sqrt{2\pi}\sigma_Z} \exp \left[ - { (z-(\mu_X+\mu_Y))^2 \over 2\sigma_Z^2 } \right]
Using the [[convolution theorem]]
It can be shown that the Fourier transform of a Gaussian, f_X(x) = \mathcal{N}(x; \mu_X, \sigma_X^2), is
:\mathcal{F}{f_X} = F_X(\omega) = \exp \left[ -j \omega \mu_X \right] \exp \left[-\tfrac{\sigma_X^2\omega^2}{2} \right]
By the convolution theorem:
: \begin{align} f_Z(z) &= (f_X * f_Y)(z) \[5pt] &= \mathcal{F}^{-1}\big{\mathcal{F}{f_X}\cdot\mathcal{F}{f_Y}\big} \[5pt] &= \mathcal{F}^{-1}\big{ \exp \left[ -j \omega \mu_X \right] \exp \left[-\tfrac{\sigma_X^2\omega^2}{2} \right] \exp \left[ -j \omega \mu_Y \right] \exp \left[-\tfrac{\sigma_Y^2\omega^2}{2} \right] \big} \[5pt] &= \mathcal{F}^{-1}\big{ \exp \left[ -j \omega (\mu_X + \mu_Y) \right] \exp \left[-\tfrac{(\sigma_X^2\ + \sigma_Y^2)\omega^2}{2} \right] \big} \[5pt] &= \mathcal{N}(z; \mu_X + \mu_Y, \sigma_X^2 + \sigma_Y^2) \end{align}
Geometric proof
First consider the normalized case when X, Y ~ N(0, 1), so that their PDFs are :f(x) = \frac 1 {\sqrt{2\pi ,}} e^{-x^2/2} and :g(y) = \frac1 {\sqrt{2\pi,}} e^{-y^2/2}. Let Z = X + Y. Then the CDF for Z will be :z \mapsto \int_{x+y \leq z} f(x)g(y) , dx , dy. This integral is over the half-plane which lies under the line x+y = z.
The key observation is that the function
: f(x)g(y) = \frac 1 {2\pi}e^{-(x^2 + y^2)/2},
is radially symmetric. So we rotate the coordinate plane about the origin, choosing new coordinates x',y' such that the line x+y = z is described by the equation x' = c where c = c(z) is determined geometrically. Because of the radial symmetry, we have f(x)g(y) = f(x')g(y') , and the CDF for Z is
:\int_{x'\leq c, y' \in \reals} f(x')g(y') , dx' , dy'.
This is easy to integrate; we find that the CDF for Z is
:\int_{-\infty}^{c(z)} f(x') , dx' = \Phi(c(z)).
To determine the value c(z), note that we rotated the plane so that the line x+y = z now runs vertically with x-intercept equal to c. So c is just the distance from the origin to the line x+y = z along the perpendicular bisector, which meets the line at its nearest point to the origin, in this case (z/2,z/2),. So the distance is c = \sqrt{ (z/2)^2 + (z/2)^2 } = z/\sqrt{2},, and the CDF for Z is \Phi(z/\sqrt{2}), i.e., Z = X+Y \sim N(0, 2).
Now, if a, b are any real constants (not both zero) then the probability that aX+bY \leq z is found by the same integral as above, but with the bounding line ax+by =z. The same rotation method works, and in this more general case we find that the closest point on the line to the origin is located a (signed) distance : \frac{z}{\sqrt{a^2 + b^2}} away, so that :aX + bY \sim N(0, a^2 + b^2). The same argument in higher dimensions shows that if :X_i \sim N(0,\sigma_i^2), \qquad i=1, \dots, n, then :X_1+ \cdots + X_n \sim N(0, \sigma_1^2 + \cdots + \sigma_n^2).
Now we are essentially done, because :X \sim N(\mu,\sigma^2) \Leftrightarrow \frac{1}{\sigma} (X - \mu) \sim N(0,1). So in general, if :X_i \sim N(\mu_i, \sigma_i^2), \qquad i=1, \dots, n, then : \sum_{i=1}^n a_i X_i \sim N\left(\sum_{i=1}^n a_i \mu_i, \sum_{i=1}^n (a_i \sigma_i)^2 \right).
References
References
- Lemons (2002) pp. 35–36
- (October 20, 2005). "Fourier Transform of the Gaussian".
This article was imported from Wikipedia and is available under the Creative Commons Attribution-ShareAlike 4.0 License. Content has been adapted to SurfDoc format. Original contributors can be found on the article history page.
Ask Mako anything about Sum of normally distributed random variables — get instant answers, deeper analysis, and related topics.
Research with MakoFree with your Surf account
Create a free account to save articles, ask Mako questions, and organize your research.
Sign up freeThis content may have been generated or modified by AI. CloudSurf Software LLC is not responsible for the accuracy, completeness, or reliability of AI-generated content. Always verify important information from primary sources.
Report