7.3: Probability from Gaussian distribution
- Page ID
- 122623
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\dsum}{\displaystyle\sum\limits} \)
\( \newcommand{\dint}{\displaystyle\int\limits} \)
\( \newcommand{\dlim}{\displaystyle\lim\limits} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)- Using standard probability density functions
- Some unique forms of pdfs (sec 16.3)
- Bernoulli -- only two f possible outcomes, but probability can be unequal
- Binomial -- multiple outcomes are possible, but distribution is uniform
- Poisson -- number of event opportunities becomes so large that likelihood is extremely small (i.e. earthquake \(\rightarrow\) tidal wave \(\rightarrow\) coastal nuclear power plant \(\rightarrow\) sea wall that is too short \ldots)
- Weibull -- time to failure of component (i.e. 99.999999% success of devices on spacecraft)
- Normal (Gaussian) -- standard bell curve
- Student t -- like Normal, but with finite number of data; sample instead of continuous population
- Normal distribution function (also referred to as Gaussian)
- If assuming an infinite number of possibilities (\(-\infty < x < \infty\) AND infinite number of decimal points), then function is \[ p(x) = \frac{1}{\sigma \sqrt{2\pi}}\mathrm{e}^{\frac{-(x-x')^{2}}{2\sigma^{2}}} \] where \(x'\) is the true mean, \(\sigma\) is the standard deviation, and \(\pi\) is an irrational but cool number NOT always related to circles
- Sketch is a bell curve with 1st central moment around \(x'\)
- Follows Central Limit Theorem: when a random process results from the summation of many elementary random processes, the result tends toward a Gaussian distribution. Examples include:
- Noise of raindrops on a roof
- Size distribution of grains on sandpaper
- Static from analog radio/TV signals
- Using a normalized Gaussian distribution
- Clarify normalized relative to normal distribution: like dividing \(p(x)\) by the magnitude to get a unit vector
- Instead of using units of \(x\), define the variable as number of standard deviations \(x\) is from the mean \[ \beta = \frac{x-x'}{\sigma} \]
- Seeking a specific value through \(z\) substitution: \[ z_{1} = \frac{x_{1}-x'}{\sigma} \]
- Calculate the probability of \(x\) from the functional form of the probability density function (i.e. integrate \(p(x)\)) \[ P(x'-\delta x \leq x \leq x'+\delta x) = \int_{x'-\delta x}^{x'+\delta x} p(x) dx \]
- Instead of using range \(\delta x\), just define as specific values: \(+x_{1} = x'+\delta x\) and \(-x_{1}=x'-\delta x\) \[ P(- x_{1} \leq x \leq +x_{1}) = \int_{-x_{1}}^{+x_{1}} p(x) dx \]
- But can also use the \(\beta\) and \(z_{1}\) if solving for the \(x\) terms \[x = \beta \sigma + x' \] \[ x_{1} = z_{1}\sigma + x' \] \[ dx = \sigma d\beta \] \[ P(-z_{1}\sigma + x' \leq \beta\sigma+x' \leq z_{1}\sigma + x') = \int_{-z_{1}\sigma + x'}^{z_{1}\sigma + x'} \frac{1}{\sigma \sqrt{2\pi}}\mathrm{e}^{\frac{-\beta^{2}}{2}}\sigma d\beta \]
- Since \(x'\) everywhere as a shift, it can be removed
- Additionally, there is a symmetry to the function to make it one sided \[ P(0\leq \beta \leq z_{1}) = 2\int_{0}^{z_{1}} \frac{1}{\sqrt{2\pi}}\mathrm{e}^{-\beta^{2}/2} d\beta \] where integral is calculated as the symmetry of the cumulative distribution function (normcdf call in Matlab)
- Tables of values for integration to standard deviations below and above the mean
- Conversion also possible through the similarity to error function \[ \frac{1}{\sqrt{2\pi}}\int_{-\infty}^{z_{1}} \mathrm{e}^{-\beta^{2}/2} d\beta = \frac{1}{2}\left[1+\mathrm{erf}\left(\frac{z_{1}}{\sqrt{2}}\right)\right] \] where the integral output is an available function call in Matlab (erf)
- To make single sided, half of the distribution is considered \[\frac{1}{\sqrt{2\pi}}\int_{0}^{z_{1}} \mathrm{e}^{-\beta^{2}/2} d\beta = \frac{1}{2}\left[\mathrm{erf}\left(\frac{z_{1}}{\sqrt{2}}\right)\right] \]
- Historic resource: table 17.2 on page 364 of Dunn & Davis textbook provides values of \(P(z_{1}) = normcdf(z_{1},0,1) - 0.5\)
- Example:
Electronic capacitors have Gaussian distribution with mean of 145 \(\mu\)F and standard deviation of 33 \(\mu\)F. The compute the probability that the capacitance is:
- greater than 189 \(\mu\)F
- less than 120 \(\mu\)F
- between 121 and 167 \(\mu\)F
- Answer
-
For the first scenario \(z_{1} = \frac{189-145}{33} = \frac{4}{3} \). Since seeking probability greater than value, the complementary cumulative is needed. From the table lookup at 1.33 decimal precision, probably is 9.17%.
Scenario #2: \(z_{1} = \frac{120-145}{33} = \frac{-25}{33} = -0.76 \). The cumulative integration less than -0.76 standard deviations is 22.36%.
The third scenario has a top and bottom boundary for the integration
- \(z_{low} = \frac{121-145}{33} = \frac{-8}{11} = -0.73 \) for the left boundary
- \(z_{high} = \frac{167-145}{33} = \frac{2}{3} = 0.67 \) for the right boundary
Multiple approaches could be considered:
- Using one-sided cumulative boundary function (\(\int_{0}^{z^{*}} p(x) dx\)) the absolute values of the standard deviations can be used: \(| z_{low}| \rightarrow 26.73\%\) probability and \(| z_{high}| \rightarrow 24.86\%\) for a total of 51.59% probability
- The difference of the full integration can be done: (\(\int_{-\infty}^{z_{high}} - \int_{-\infty}^{z_{low}}\)) where the lookup tables would have 74.86 - 23.27 = 51.59%
- Use the edges of the integration subtracted from unity (\(1 - \int_{-\infty}^{z_{low}} - \int_{z_{high}}^{\infty}\)) where the last term is the complementary cumulative: 100 - 23.27 - 25.14 = 51.59%
- Some unique forms of pdfs (sec 16.3)

