This article needs additional citations for verification. Please help improve this article by adding citations to reliable sources. Unsourced material may be challenged and removed.Find sources: "Half-normal distribution" – news · newspapers · books · scholar · JSTOR (November 2020) (Learn how and when to remove this template message)
Parameters Probability density function${\displaystyle \sigma =1}$ Cumulative distribution function${\displaystyle \sigma =1}$ ${\displaystyle \sigma >0}$ — (scale) ${\displaystyle x\in [0,\infty )}$ ${\displaystyle f(x;\sigma )={\frac {\sqrt {2)){\sigma {\sqrt {\pi ))))\exp \left(-{\frac {x^{2)){2\sigma ^{2))}\right)\quad x>0}$ ${\displaystyle F(x;\sigma )=\operatorname {erf} \left({\frac {x}{\sigma {\sqrt {2))))\right)}$ ${\displaystyle Q(F;\sigma )=\sigma {\sqrt {2))\operatorname {erf} ^{-1}(F)}$ ${\displaystyle {\frac {\sigma {\sqrt {2))}{\sqrt {\pi ))))$ ${\displaystyle \sigma {\sqrt {2))\operatorname {erf} ^{-1}(1/2)}$ ${\displaystyle 0}$ ${\displaystyle \sigma ^{2}\left(1-{\frac {2}{\pi ))\right)}$ ${\displaystyle {\frac ((\sqrt {2))(4-\pi )}{(\pi -2)^{3/2))}\approx 0.9952717}$ ${\displaystyle {\frac {8(\pi -3)}{(\pi -2)^{2))}\approx 0.869177}$ ${\displaystyle {\frac {1}{2))\log _{2}\left(2\pi e\sigma ^{2}\right)-1}$

In probability theory and statistics, the half-normal distribution is a special case of the folded normal distribution.

Let ${\displaystyle X}$ follow an ordinary normal distribution, ${\displaystyle N(0,\sigma ^{2})}$. Then, ${\displaystyle Y=|X|}$ follows a half-normal distribution. Thus, the half-normal distribution is a fold at the mean of an ordinary normal distribution with mean zero.

## Properties

Using the ${\displaystyle \sigma }$ parametrization of the normal distribution, the probability density function (PDF) of the half-normal is given by

${\displaystyle f_{Y}(y;\sigma )={\frac {\sqrt {2)){\sigma {\sqrt {\pi ))))\exp \left(-{\frac {y^{2)){2\sigma ^{2))}\right)\quad y\geq 0,}$

where ${\displaystyle E[Y]=\mu ={\frac {\sigma {\sqrt {2))}{\sqrt {\pi ))))$.

Alternatively using a scaled precision (inverse of the variance) parametrization (to avoid issues if ${\displaystyle \sigma }$ is near zero), obtained by setting ${\displaystyle \theta ={\frac {\sqrt {\pi )){\sigma {\sqrt {2))))}$, the probability density function is given by

${\displaystyle f_{Y}(y;\theta )={\frac {2\theta }{\pi ))\exp \left(-{\frac {y^{2}\theta ^{2)){\pi ))\right)\quad y\geq 0,}$

where ${\displaystyle E[Y]=\mu ={\frac {1}{\theta ))}$.

The cumulative distribution function (CDF) is given by

${\displaystyle F_{Y}(y;\sigma )=\int _{0}^{y}{\frac {1}{\sigma )){\sqrt {\frac {2}{\pi ))}\,\exp \left(-{\frac {x^{2)){2\sigma ^{2))}\right)\,dx}$

Using the change-of-variables ${\displaystyle z=x/({\sqrt {2))\sigma )}$, the CDF can be written as

${\displaystyle F_{Y}(y;\sigma )={\frac {2}{\sqrt {\pi ))}\,\int _{0}^{y/({\sqrt {2))\sigma )}\exp \left(-z^{2}\right)dz=\operatorname {erf} \left({\frac {y}((\sqrt {2))\sigma ))\right),}$

where erf is the error function, a standard function in many mathematical software packages.

The quantile function (or inverse CDF) is written:

${\displaystyle Q(F;\sigma )=\sigma {\sqrt {2))\operatorname {erf} ^{-1}(F)}$

where ${\displaystyle 0\leq F\leq 1}$ and ${\displaystyle \operatorname {erf} ^{-1))$ is the inverse error function

The expectation is then given by

${\displaystyle E[Y]=\sigma {\sqrt {2/\pi )),}$

The variance is given by

${\displaystyle \operatorname {var} (Y)=\sigma ^{2}\left(1-{\frac {2}{\pi ))\right).}$

Since this is proportional to the variance σ2 of X, σ can be seen as a scale parameter of the new distribution.

The differential entropy of the half-normal distribution is exactly one bit less the differential entropy of a zero-mean normal distribution with the same second moment about 0. This can be understood intuitively since the magnitude operator reduces information by one bit (if the probability distribution at its input is even). Alternatively, since a half-normal distribution is always positive, the one bit it would take to record whether a standard normal random variable were positive (say, a 1) or negative (say, a 0) is no longer necessary. Thus,

${\displaystyle h(Y)={\frac {1}{2))\log _{2}\left({\frac {\pi e\sigma ^{2)){2))\right)={\frac {1}{2))\log _{2}\left(2\pi e\sigma ^{2}\right)-1.}$

## Applications

The half-normal distribution is commonly utilized as a prior probability distribution for variance parameters in Bayesian inference applications.[1][2]

## Parameter estimation

Given numbers ${\displaystyle \{x_{i}\}_{i=1}^{n))$ drawn from a half-normal distribution, the unknown parameter ${\displaystyle \sigma }$ of that distribution can be estimated by the method of maximum likelihood, giving

${\displaystyle {\hat {\sigma ))={\sqrt ((\frac {1}{n))\sum _{i=1}^{n}x_{i}^{2))))$

The bias is equal to

${\displaystyle b\equiv \operatorname {E} {\bigg [}\;({\hat {\sigma ))_{\mathrm {mle} }-\sigma )\;{\bigg ]}=-{\frac {\sigma }{4n))}$

which yields the bias-corrected maximum likelihood estimator

${\displaystyle {\hat {\sigma \,))_{\text{mle))^{*}={\hat {\sigma \,))_{\text{mle))-{\hat {b\,)).}$

## Related distributions

• The distribution is a special case of the folded normal distribution with μ = 0.
• It also coincides with a zero-mean normal distribution truncated from below at zero (see truncated normal distribution)
• If Y has a half-normal distribution, then (Y/σ)2 has a chi square distribution with 1 degree of freedom, i.e. Y/σ has a chi distribution with 1 degree of freedom.
• The half-normal distribution is a special case of the generalized gamma distribution with d = 1, p = 2, a = ${\displaystyle {\sqrt {2))\sigma }$.
• If Y has a half-normal distribution, Y -2 has a Levy distribution
• The Rayleigh distribution is a moment-tilted and scaled generalization of the half-normal distribution.
• Modified half-normal distribution[3] with the pdf on ${\displaystyle (0,\infty )}$ is given as ${\displaystyle f(x)={\frac {2\beta ^{\frac {\alpha }{2))x^{\alpha -1}\exp(-\beta x^{2}+\gamma x)}{\Psi {\left({\frac {\alpha }{2)),{\frac {\gamma }{\sqrt {\beta ))}\right)))))$, where ${\displaystyle \Psi (\alpha ,z)={}_{1}\Psi _{1}\left({\begin{matrix}\left(\alpha ,{\frac {1}{2))\right)\\(1,0)\end{matrix));z\right)}$ denotes the Fox–Wright Psi function.

## References

1. ^ Gelman, A. (2006), "Prior distributions for variance parameters in hierarchical models", Bayesian Analysis, 1 (3): 515–534, doi:10.1214/06-ba117a
2. ^ Röver, C.; Bender, R.; Dias, S.; Schmid, C.H.; Schmidli, H.; Sturtz, S.; Weber, S.; Friede, T. (2021), "On weakly informative prior distributions for the heterogeneity parameter in Bayesian random‐effects meta‐analysis", Research Synthesis Methods, 12 (4): 448–474, arXiv:2007.08352, doi:10.1002/jrsm.1475, PMID 33486828, S2CID 220546288
3. ^ Sun, Jingchao; Kong, Maiying; Pal, Subhadip (22 June 2021). "The Modified-Half-Normal distribution: Properties and an efficient sampling scheme". Communications in Statistics - Theory and Methods. 52 (5): 1591–1613. doi:10.1080/03610926.2021.1934700. ISSN 0361-0926. S2CID 237919587.

## Further reading

(note that MathWorld uses the parameter ${\displaystyle \theta ={\frac {1}{\sigma )){\sqrt {\pi /2))}$