请输入您要查询的百科知识:

 

词条 Logit
释义

  1. Definition

  2. History

  3. Uses and properties

  4. Comparison with probit

  5. See also

  6. References

  7. Further reading

{{Hatnote|This article discusses the binary logit function only. See discrete choice for a discussion of multinomial logit, conditional logit, nested logit, mixed logit, exploded logit, and ordered logit. For the basic regression technique that uses the logit function, see logistic regression.}}

In statistics, the logit ({{IPAc-en|ˈ|l|oʊ|dʒ|ɪ|t}} {{respell|LOH|jit}}) function or the log-odds is the logarithm of the odds {{math|p/(1 − p)}} where {{mvar|p}} is the probability.[1]. It is a type of function that creates a map of probability values from to [2]. It is the inverse of the sigmoidal "logistic" function or logistic transform used in mathematics, especially in statistics.

In deep learning, the term logits layer is popularly used for the last neuron layer of neural network for classification task which produces raw prediction values as real numbers ranging from [3].

Definition

If p is a probability, then {{nowrap|p/(1 − p)}} is the corresponding odds; the logit of the probability is the logarithm of the odds, i.e.

The base of the logarithm function used is of little importance in the present article, as long as it is greater than 1, but the natural logarithm with base e is the one most often used. The choice of base corresponds to the choice of logarithmic unit for the value: base 2 corresponds to a shannon, base e to a nat, and base 10 to a hartley; these units are particularly used in information-theoretic interpretations. For each choice of base, the logit function takes values between negative and positive infinity.

The "logistic" function of any number is given by the inverse-logit:

The difference between the logits of two probabilities is the logarithm of the odds ratio (R), thus providing a shorthand for writing the correct combination of odds ratios only by adding and subtracting:

History

There have been several efforts to adapt linear regression methods to domain where output is probability value instead of any real number . Many of such efforts focused on modeling this problem by somehow mapping the range to and then running the linear regression on these transformed values. In 1934 Chester Ittner Bliss used the cumulative normal distribution function to perform this mapping and called his model probit an abbreviation for "probability unit";[4] . However, this is computationally more expensive. In 1944, Joseph Berkson used log of odds and called this function logit, abbreviation for "logistic unit" following the analogy for probit. Log odds was used extensively by Charles Sanders Peirce (late 19th century).[5] . G. A. Barnard in 1949 coined the commonly used term log-odds;[6] the log-odds of an event is the logit of the probability of the event.[7]

Uses and properties

  • The logit in logistic regression is a special case of a link function in a generalized linear model: it is the canonical link function for the Bernoulli distribution.
  • The logit function is the negative of the derivative of the binary entropy function.
  • The logit is also central to the probabilistic Rasch model for measurement, which has applications in psychological and educational assessment, among other areas.
  • The inverse-logit function (i.e., the logistic function) is also sometimes referred to as the expit function.[8]
  • In plant disease epidemiology the logit is used to fit the data to a logistic model. With the Gompertz and Monomolecular models all three are known as Richards family models.
  • The log-odds function of probabilities is often used in state estimation algorithms[9] because of its numerical advantages in the case of small probabilities. Instead of multiplying very small floating point numbers, log-odds probabilities can just be summed up to calculate the (log-odds) joint probability.[10][11]

Comparison with probit

Closely related to the logit function (and logit model) are the probit function and probit model. The logit and probit are both sigmoid functions with a domain between 0 and 1, which makes them both quantile functions—i.e., inverses of the cumulative distribution function (CDF) of a probability distribution. In fact, the logit is the quantile function of the logistic distribution, while the probit is the quantile function of the normal distribution. The probit function is denoted , where is the CDF of the normal distribution, as just mentioned:

As shown in the graph, the logit and probit functions are extremely similar, particularly when the probit function is scaled so that its slope at y=0 matches the slope of the logit. As a result, probit models are sometimes used in place of logit models because for certain applications (e.g., in Bayesian statistics) the implementation is easier.

See also

  • Discrete choice on binary logit, multinomial logit, conditional logit, nested logit, mixed logit, exploded logit, and ordered logit
  • Limited dependent variable
  • Daniel McFadden, a Nobel Prize in Economics winner for development of a particular logit model used in economics[4]
  • Logit analysis in marketing
  • Multinomial logit
  • Ogee, curve with similar shape
  • Perceptron
  • Probit, another function with the same domain and range as the logit
  • Ridit scoring
  • Data transformation (statistics)
  • Arcsin (transformation)

References

{{More footnotes|date=November 2010}}
1. ^{{cite web|url=http://itl.nist.gov/div898/software/dataplot/refman2/auxillar/logoddra.htm|title=LOG ODDS RATIO|work=nist.gov}}
2. ^{{Cite web|url=http://www.columbia.edu/~so33/SusDev/Lecture_9.pdf|title=Logit/Probit|last=|first=|date=|website=|archive-url=|archive-date=|dead-url=|access-date=}}
3. ^{{Cite web|url=https://www.tensorflow.org/tutorials/estimators/cnn#logits_layer|title=Tensorflow - Logit Layer|last=|first=|date=|website=|archive-url=|archive-date=|dead-url=|access-date=}}
4. ^{{Cite web|url=http://www.cambridge.org/resources/0521815886/1208_default.pdf|title=The origins and development of the logit model|author=J. S. Cramer|year=2003|publisher=Cambridge UP}}
5. ^{{cite book |title=The history of statistics : the measurement of uncertainty before 1900 |last=Stigler |first=Stephen M. |authorlink=Stephen M. Stigler |year=1986 |publisher=Belknap Press of Harvard University Press |location=Cambridge, Massachusetts |isbn=978-0-674-40340-6 |ref=harv }}
6. ^{{citation|title=Logistic Regression Models|first=Joseph M.|last=Hilbe|publisher=CRC Press|year=2009|isbn=9781420075779|page=3|url=https://books.google.com/books?id=tmHMBQAAQBAJ&pg=PA3}}.
7. ^{{citation|title=Logit Models from Economics and Other Fields|first=J. S.|last=Cramer|publisher=Cambridge University Press|year=2003|isbn=9781139438193|page=13|url=https://books.google.com/books?id=1Od2d72pPXUC&pg=PA13}}.
8. ^{{cite web |url=http://www.stat.ucl.ac.be/ISdidactique/Rhelp/library/msm/html/expit.html |title=Archived copy |accessdate=2011-02-18 |deadurl=yes |archiveurl=https://web.archive.org/web/20110706132209/http://www.stat.ucl.ac.be/ISdidactique/Rhelp/library/msm/html/expit.html |archivedate=2011-07-06 |df= }}
9. ^{{Cite journal|last=Thrun|first=Sebastian|title=Learning Occupancy Grid Maps with Forward Sensor Models|journal=Autonomous Robots|language=en|volume=15|issue=2|pages=111–127|doi=10.1023/A:1025584807625|issn=0929-5593|year=2003}}
10. ^{{Cite web|url=http://www.cs.cmu.edu/~16831-f12/notes/F12/16831_lecture05_vh.pdf|title=Statistical Techniques in Robotics|last=Styler|first=Alex|date=2012|website=|publisher=|page=2|access-date=2017-01-26}}
11. ^{{Cite journal|last=Dickmann|first=J.|last2=Appenrodt|first2=N.|last3=Klappstein|first3=J.|last4=Bloecher|first4=H. L.|last5=Muntzinger|first5=M.|last6=Sailer|first6=A.|last7=Hahn|first7=M.|last8=Brenk|first8=C.|date=2015-01-01|title=Making Bertha See Even More: Radar Contribution|url=http://ieeexplore.ieee.org/document/7161279/|journal=IEEE Access|volume=3|pages=1233–1247|doi=10.1109/ACCESS.2015.2454533|issn=2169-3536}}

Further reading

  • {{cite book|last=Ashton|first=Winifred D.|title=The Logit Transformation: with special reference to its uses in Bioassay|year=1972|publisher=Charles Griffin|isbn=978-0-85264-212-2|series=Griffin's Statistical Monographs & Courses|volume= 32}}

2 : Logarithms|Special functions

随便看

 

开放百科全书收录14589846条英语、德语、日语等多语种百科知识,基本涵盖了大多数领域的百科知识,是一部内容自由、开放的电子版国际百科全书。

 

Copyright © 2023 OENC.NET All Rights Reserved
京ICP备2021023879号 更新时间:2024/11/18 14:07:42