World Library  
Flag as Inappropriate
Email this Article

Logit-normal distribution

Article Id: WHEBN0028843342
Reproduction Date:

Title: Logit-normal distribution  
Author: World Heritage Encyclopedia
Language: English
Subject: Probability Distribution, Multivariate Pareto distribution, Hyper-Erlang distribution, Dyadic distribution, Chernoff's distribution
Collection: Continuous Distributions, Probability Distributions
Publisher: World Heritage Encyclopedia
Publication
Date:
 

Logit-normal distribution

Logit-normal
Probability density function
Plot of the Logitnormal PDF
Cumulative distribution function
Plot of the Logitnormal PDF
Notation P( \mathcal{N}(\mu,\,\sigma^2) )
Parameters σ2 > 0 — squared scale (real),
μR — location
Support x ∈ (0, 1)
pdf \frac{1}{\sigma \sqrt{2 \pi}}\, e^{-\frac{(\operatorname{logit}(x) - \mu)^2}{2\sigma^2}}\frac{1}{x (1-x)}
CDF \frac12\Big[1 + \operatorname{erf}\Big( \frac{\operatorname{logit}(x)-\mu}{\sqrt{2\sigma^2}}\Big)\Big]
Mean no analytical solution
Median P(\mu)\,
Mode no analytical solution
Variance no analytical solution
MGF no analytical solution

In probability theory, a logit-normal distribution is a probability distribution of a random variable whose logit has a normal distribution. If Y is a random variable with a normal distribution, and P is the logistic function, then X = P(Y) has a logit-normal distribution; likewise, if X is logit-normally distributed, then Y = logit(X)= log (X/(1-X)) is normally distributed. It is also known as the logistic normal distribution,[1] which often refers to a multinomial logit version (e.g.[2][3][4][5]).

A variable might be modeled as logit-normal if it is a proportion, which is bounded by zero and one, and where values of zero and one never occur.

Contents

  • Characterization 1
    • Probability density function 1.1
    • Moments 1.2
    • Mode 1.3
  • Multivariate generalization 2
    • Probability density function 2.1
    • Use in statistical analysis 2.2
    • Relationship with the Dirichlet distribution 2.3
  • See also 3
  • Further reading 4
  • External links 5

Characterization

Probability density function

The probability density function (PDF) of a logit-normal distribution, for 0 ≤ x ≤ 1, is:

f_X(x;\mu,\sigma) = \frac{1}{\sigma\sqrt{2 \pi}}\,\frac{1}{x (1-x)}\, e^{-\frac{(\operatorname{logit}(x) - \mu)^2}{2\sigma^2}}

where μ and σ are the mean and standard deviation of the variable’s logit (by definition, the variable’s logit is normally distributed).

The density obtained by changing the sign of μ is symmetrical, in that it is equal to f(1-x;-μ,σ), shifting the mode to the other side of 0.5 (the midpoint of the (0,1) interval).

Plot of the Logitnormal PDF for various combinations of μ (facets) and σ (colors)

Moments

The moments of the logit-normal distribution have no analytic solution. However, they can be estimated by numerical integration.

Mode

When the derivative of the density equals 0 then the location of the mode x satisfies the following equation:

\operatorname{logit}(x) = \sigma^2(2x-1)+\mu .

Multivariate generalization

The logistic normal distribution is a generalization of the logit–normal distribution to D-dimensional probability vectors by taking a logistic transformation of a multivariate normal distribution.[6][7][8]

Probability density function

The probability density function is:

f_X( \mathbf{x}; \boldsymbol{\mu} , \boldsymbol{\Sigma} ) = \frac{1}{ | 2 \pi \boldsymbol{\Sigma} |^\frac{1}{2} } \, \frac{1}{ \prod\limits_{i=1}^D x_i } \, e^{- \frac{1}{2} \left\{ \log \left( \frac{ \mathbf{x}_{-D} }{ x_D } \right) - \boldsymbol{\mu} \right\}^\top \boldsymbol{\Sigma}^{-1} \left\{ \log \left( \frac{ \mathbf{x}_{-D} }{ x_D } \right) - \boldsymbol{\mu} \right\} } \quad , \quad \mathbf{x} \in \mathcal{S}^D \;\; ,

where \mathbf{x}_{-D} denotes a vector of the first (D-1) components of \mathbf{x} and \mathcal{S}^D denotes the simplex of D-dimensional probability vectors. This follows from applying the additive logistic transformation to map a multivariate normal random variable \mathbf{y} \sim \mathcal{N} \left( \boldsymbol{\mu} , \boldsymbol{\Sigma} \right) \; , \; \mathbf{y} \in \mathbb{R}^{D-1} to the simplex:

\mathbf{x} = \left[ \frac{ e^{ y_1 } }{ 1 + \sum_{i=1}^{D-1} e^{ y_i } } , \dots , \frac{ e^{ y_{D-1} } }{ 1 + \sum_{i=1}^{D-1} e^{ y_i } } , \frac{ 1 }{ 1 + \sum_{i=1}^{D-1} e^{ y_i } } \right]
Gaussian density functions and corresponding logistic normal density functions after logistic transformation.

The unique inverse mapping is given by:

\mathbf{y} = \left[ \log \left( \frac{ x_1 }{ x_D } \right) , \dots , \log \left( \frac{ x_{D-1} }{ x_D } \right) \right]

Use in statistical analysis

The logistic normal distribution is a more flexible alternative to the Dirichlet distribution in that it can capture correlations between components of probability vectors. It also has the potential to simplify statistical analyses of compositional data by allowing one to answer questions about log-ratios of the components of the data vectors. One is often interested in ratios rather than absolute component values.

The probability simplex is a bounded space, making standard techniques that are typically applied to vectors in \mathbb{R}^n less meaningful. Aitchison described the problem of spurious negative correlations when applying such methods directly to simplicial vectors.[7] However, mapping compositional data in \mathcal{S}^D through the inverse of the additive logistic transformation yields real-valued data in \mathbb{R}^{D-1}. Standard techniques can be applied to this representation of the data. This approach justifies use of the logistic normal distribution, which can thus be regarded as the "Gaussian of the simplex".

Relationship with the Dirichlet distribution

Logistic normal approximation to Dirichlet distribution

The Dirichlet and logistic normal distributions are never exactly equal for any choice of parameters. However, Aitchison described a method for approximating a Dirichlet with a logistic normal such that their Kullback–Leibler divergence (KL) is minimized:

K(p,q) = \int_{\mathcal{S}^D} p \left( \mathbf{x} | \boldsymbol{\alpha} \right) \log \left( \frac{ p \left( \mathbf{x} | \boldsymbol{\alpha} \right) }{ q \left( \mathbf{x} | \boldsymbol{\mu} , \boldsymbol{\Sigma} \right) } \right) \, d \mathbf{x}

This is minimized by:

\boldsymbol{\mu}^* = \mathbf{E}_p \left[ \log \left( \frac{ \mathbf{x}_{-D} }{ x_D } \right) \right] \quad , \quad \boldsymbol{\Sigma}^* = \textbf{Var}_p \left[ \log \left( \frac{ \mathbf{x}_{-D} }{ x_D } \right) \right]

Using moment properties of the Dirichlet distribution, the solution can be written in terms of the digamma \psi and trigamma \psi' functions:

\mu_i^* = \psi \left( \alpha_i \right) - \psi \left( \alpha_D \right) \quad , \quad i = 1 , \cdots , D-1
\Sigma_{ii}^* = \psi' \left( \alpha_i \right) + \psi' \left( \alpha_D \right) \quad , \quad i = 1 , \cdots , D-1
\Sigma_{ij}^* = \psi' \left( \alpha_D \right) \quad , \quad i \neq j

This approximation is particularly accurate for large \boldsymbol{\alpha}. In fact, one can show that for \alpha_i \rightarrow \infty , i = 1 , \cdots , D, we have that p \left( \mathbf{x} | \boldsymbol{\alpha} \right) \rightarrow q \left( \mathbf{x} | \boldsymbol{\mu}^* , \boldsymbol{\Sigma}^* \right).

See also

Further reading

  • Frederic, P. & Lad, F. (2008) Two Moments of the Logitnormal Distribution. Communications in Statistics-Simulation and Computation. 37: 1263-1269
  • Mead, R. (1965). "A Generalised Logit-Normal Distribution". Biometrics 21 (3): 721–732.  
  1. ^ J Atchison and SM Shen. "Logistic-normal distributions: Some properties and uses." Biometrika, 1980. Google Scholar link
  2. ^ http://people.csail.mit.edu/tomasz/papers/huang_hln_tech_report_2006.pdf
  3. ^ Peter Hoff, 2003. Link
  4. ^ http://www.springerreference.com/docs/html/chapterdbid/205424.html
  5. ^ http://brenocon.com/blog/2011/05/log-normal-and-logistic-normal-terminology/
  6. ^ Aitchison, J.; Shen, S. M. (1980). "Logistic-normal distributions: Some properties and uses". Biometrika 67 (2): 261.  
  7. ^ a b J. Atchison. "The Statistical Analysis of Compositional Data." Monographs on Statistics and Applied Probability, Chapman and Hall, 1986. Book
  8. ^  

External links

  • logitnorm package for R
This article was sourced from Creative Commons Attribution-ShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and USA.gov, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for USA.gov and content contributors is made possible from the U.S. Congress, E-Government Act of 2002.
 
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
 
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a non-profit organization.
 


Copyright © World Library Foundation. All rights reserved. eBooks from Project Gutenberg are sponsored by the World Library Foundation,
a 501c(4) Member's Support Non-Profit Organization, and is NOT affiliated with any governmental agency or department.