Logistic regression, the Glossary
In statistics, the logistic model (or logit model) is a statistical model that models the log-odds of an event as a linear combination of one or more independent variables.[1]
Table of Contents
190 relations: Adolphe Quetelet, Analytic function, Autocatalysis, Backpropagation, Bayesian statistics, Bernoulli distribution, Bernoulli trial, Binary classification, Binary data, Binary regression, Binomial distribution, Bioassay, Blood test, Body mass index, Brier score, Built environment, C (programming language), C++, Cambridge University Press, Canada, Canonical form, Cardinal number, Catalysis, Categorical variable, Chester Ittner Bliss, Chi-squared test, Coefficient of determination, Computer science, Conditional entropy, Conditional logistic regression, Conditional random field, Conjugate prior, Contingency table, Continuous or discrete variable, Coronary artery disease, Cross-entropy, Cumulative distribution function, Daniel McFadden, David Cox (statistician), Degrees of freedom (statistics), Dependent and independent variables, Design matrix, Deviance (statistics), Diabetes, Discrete choice, Dot product, Dummy variable (statistics), E (mathematical constant), Econometrics, Economics, ... Expand index (140 more) »
- Predictive analytics
- Regression models
Adolphe Quetelet
Lambert Adolphe Jacques Quetelet FRSF or FRSE (22 February 1796 – 17 February 1874) was a Belgian astronomer, mathematician, statistician and sociologist who founded and directed the Brussels Observatory and was influential in introducing statistical methods to the social sciences.
See Logistic regression and Adolphe Quetelet
Analytic function
In mathematics, an analytic function is a function that is locally given by a convergent power series.
See Logistic regression and Analytic function
Autocatalysis
In chemistry, a chemical reaction is said to be autocatalytic if one of the reaction products is also a catalyst for the same reaction.
See Logistic regression and Autocatalysis
Backpropagation
In machine learning, backpropagation is a gradient estimation method used to train neural network models.
See Logistic regression and Backpropagation
Bayesian statistics
Bayesian statistics is a theory in the field of statistics based on the Bayesian interpretation of probability, where probability expresses a degree of belief in an event.
See Logistic regression and Bayesian statistics
Bernoulli distribution
In probability theory and statistics, the Bernoulli distribution, named after Swiss mathematician Jacob Bernoulli, is the discrete probability distribution of a random variable which takes the value 1 with probability p and the value 0 with probability q.
See Logistic regression and Bernoulli distribution
Bernoulli trial
In the theory of probability and statistics, a Bernoulli trial (or binomial trial) is a random experiment with exactly two possible outcomes, "success" and "failure", in which the probability of success is the same every time the experiment is conducted.
See Logistic regression and Bernoulli trial
Binary classification
Binary classification is the task of classifying the elements of a set into one of two groups (each called class).
See Logistic regression and Binary classification
Binary data
Binary data is data whose unit can take on only two possible states.
See Logistic regression and Binary data
Binary regression
In statistics, specifically regression analysis, a binary regression estimates a relationship between one or more explanatory variables and a single output binary variable.
See Logistic regression and Binary regression
Binomial distribution
In probability theory and statistics, the binomial distribution with parameters and is the discrete probability distribution of the number of successes in a sequence of independent experiments, each asking a yes–no question, and each with its own Boolean-valued outcome: success (with probability) or failure (with probability).
See Logistic regression and Binomial distribution
Bioassay
A bioassay is an analytical method to determine the potency or effect of a substance by its effect on living animals or plants (in vivo), or on living cells or tissues (in vitro).
See Logistic regression and Bioassay
Blood test
A blood test is a laboratory analysis performed on a blood sample that is usually extracted from a vein in the arm using a hypodermic needle, or via fingerprick.
See Logistic regression and Blood test
Body mass index
Body mass index (BMI) is a value derived from the mass (weight) and height of a person.
See Logistic regression and Body mass index
Brier score
The Brier Score is a ''strictly proper score function'' or ''strictly proper scoring rule'' that measures the accuracy of probabilistic predictions.
See Logistic regression and Brier score
Built environment
The term built environment refers to human-made conditions and is often used in architecture, landscape architecture, urban planning, public health, sociology, and anthropology, among others.
See Logistic regression and Built environment
C (programming language)
C (pronounced – like the letter c) is a general-purpose programming language.
See Logistic regression and C (programming language)
C++
C++ (pronounced "C plus plus" and sometimes abbreviated as CPP) is a high-level, general-purpose programming language created by Danish computer scientist Bjarne Stroustrup.
See Logistic regression and C++
Cambridge University Press
Cambridge University Press is the university press of the University of Cambridge.
See Logistic regression and Cambridge University Press
Canada
Canada is a country in North America.
See Logistic regression and Canada
Canonical form
In mathematics and computer science, a canonical, normal, or standard form of a mathematical object is a standard way of presenting that object as a mathematical expression.
See Logistic regression and Canonical form
Cardinal number
In mathematics, a cardinal number, or cardinal for short, is what is commonly called the number of elements of a set.
See Logistic regression and Cardinal number
Catalysis
Catalysis is the increase in rate of a chemical reaction due to an added substance known as a catalyst.
See Logistic regression and Catalysis
Categorical variable
In statistics, a categorical variable (also called qualitative variable) is a variable that can take on one of a limited, and usually fixed, number of possible values, assigning each individual or other unit of observation to a particular group or nominal category on the basis of some qualitative property.
See Logistic regression and Categorical variable
Chester Ittner Bliss
Chester Ittner Bliss (February 1, 1899 – March 14, 1979) was primarily a biologist, who is best known for his contributions to statistics.
See Logistic regression and Chester Ittner Bliss
Chi-squared test
A chi-squared test (also chi-square or test) is a statistical hypothesis test used in the analysis of contingency tables when the sample sizes are large.
See Logistic regression and Chi-squared test
Coefficient of determination
In statistics, the coefficient of determination, denoted R2 or r2 and pronounced "R squared", is the proportion of the variation in the dependent variable that is predictable from the independent variable(s).
See Logistic regression and Coefficient of determination
Computer science
Computer science is the study of computation, information, and automation.
See Logistic regression and Computer science
Conditional entropy
In information theory, the conditional entropy quantifies the amount of information needed to describe the outcome of a random variable Y given that the value of another random variable X is known.
See Logistic regression and Conditional entropy
Conditional logistic regression
Conditional logistic regression is an extension of logistic regression that allows one to account for stratification and matching.
See Logistic regression and Conditional logistic regression
Conditional random field
Conditional random fields (CRFs) are a class of statistical modeling methods often applied in pattern recognition and machine learning and used for structured prediction.
See Logistic regression and Conditional random field
Conjugate prior
In Bayesian probability theory, if, given a likelihood function p(x \mid \theta), the posterior distribution p(\theta \mid x) is in the same probability distribution family as the prior probability distribution p(\theta), the prior and posterior are then called conjugate distributions with respect to that likelihood function and the prior is called a conjugate prior for the likelihood function p(x \mid \theta).
See Logistic regression and Conjugate prior
Contingency table
In statistics, a contingency table (also known as a cross tabulation or crosstab) is a type of table in a matrix format that displays the multivariate frequency distribution of the variables.
See Logistic regression and Contingency table
Continuous or discrete variable
In mathematics and statistics, a quantitative variable may be continuous or discrete if they are typically obtained by measuring or counting, respectively.
See Logistic regression and Continuous or discrete variable
Coronary artery disease
Coronary artery disease (CAD), also called coronary heart disease (CHD), ischemic heart disease (IHD), myocardial ischemia, or simply heart disease, involves the reduction of blood flow to the cardiac muscle due to build-up of atherosclerotic plaque in the arteries of the heart.
See Logistic regression and Coronary artery disease
Cross-entropy
In information theory, the cross-entropy between two probability distributions p and q, over the same underlying set of events, measures the average number of bits needed to identify an event drawn from the set when the coding scheme used for the set is optimized for an estimated probability distribution q, rather than the true distribution p.
See Logistic regression and Cross-entropy
Cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Every probability distribution supported on the real numbers, discrete or "mixed" as well as continuous, is uniquely identified by a right-continuous monotone increasing function (a càdlàg function) F \colon \mathbb R \rightarrow satisfying \lim_F(x).
See Logistic regression and Cumulative distribution function
Daniel McFadden
Daniel Little McFadden (born July 29, 1937) is an American econometrician who shared the 2000 Nobel Memorial Prize in Economic Sciences with James Heckman.
See Logistic regression and Daniel McFadden
David Cox (statistician)
Sir David Roxbee Cox (15 July 1924 – 18 January 2022) was a British statistician and educator.
See Logistic regression and David Cox (statistician)
Degrees of freedom (statistics)
In statistics, the number of degrees of freedom is the number of values in the final calculation of a statistic that are free to vary.
See Logistic regression and Degrees of freedom (statistics)
Dependent and independent variables
A variable is considered dependent if it depends on an independent variable.
See Logistic regression and Dependent and independent variables
Design matrix
In statistics and in particular in regression analysis, a design matrix, also known as model matrix or regressor matrix and often denoted by X, is a matrix of values of explanatory variables of a set of objects.
See Logistic regression and Design matrix
Deviance (statistics)
In statistics, deviance is a goodness-of-fit statistic for a statistical model; it is often used for statistical hypothesis testing.
See Logistic regression and Deviance (statistics)
Diabetes
Diabetes mellitus, often known simply as diabetes, is a group of common endocrine diseases characterized by sustained high blood sugar levels.
See Logistic regression and Diabetes
Discrete choice
In economics, discrete choice models, or qualitative choice models, describe, explain, and predict choices between two or more discrete alternatives, such as entering or not entering the labor market, or choosing between modes of transport.
See Logistic regression and Discrete choice
Dot product
In mathematics, the dot product or scalar productThe term scalar product means literally "product with a scalar as a result".
See Logistic regression and Dot product
Dummy variable (statistics)
In regression analysis, a dummy variable (also known as indicator variable or just dummy) is one that takes a binary value (0 or 1) to indicate the absence or presence of some categorical effect that may be expected to shift the outcome.
See Logistic regression and Dummy variable (statistics)
E (mathematical constant)
The number is a mathematical constant approximately equal to 2.71828 that can be characterized in many ways.
See Logistic regression and E (mathematical constant)
Econometrics
Econometrics is an application of statistical methods to economic data in order to give empirical content to economic relationships.
See Logistic regression and Econometrics
Economics
Economics is a social science that studies the production, distribution, and consumption of goods and services.
See Logistic regression and Economics
Edwin Bidwell Wilson
Edwin Bidwell Wilson (April 25, 1879 – December 28, 1964) was an American mathematician, statistician, physicist and general polymath.
See Logistic regression and Edwin Bidwell Wilson
Emergency management
Emergency management (also disaster management) is a science and a system charged with creating the framework within which communities reduce vulnerability to hazards and cope with disasters.
See Logistic regression and Emergency management
Engineering
Engineering is the practice of using natural science, mathematics, and the engineering design process to solve technical problems, increase efficiency and productivity, and improve systems.
See Logistic regression and Engineering
Entropy (information theory)
In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes.
See Logistic regression and Entropy (information theory)
Error function
In mathematics, the error function (also called the Gauss error function), often denoted by, is a function defined as: \operatorname z.
See Logistic regression and Error function
Estimation theory
Estimation theory is a branch of statistics that deals with estimating the values of parameters based on measured empirical data that has a random component.
See Logistic regression and Estimation theory
Euler numbers
In mathematics, the Euler numbers are a sequence En of integers defined by the Taylor series expansion where \cosh (t) is the hyperbolic cosine function.
See Logistic regression and Euler numbers
Expectation propagation
Expectation propagation (EP) is a technique in Bayesian machine learning.
See Logistic regression and Expectation propagation
Expected value
In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first moment) is a generalization of the weighted average.
See Logistic regression and Expected value
Exponential family
In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below.
See Logistic regression and Exponential family
Exponential function
The exponential function is a mathematical function denoted by f(x).
See Logistic regression and Exponential function
Generalized linear model
In statistics, a generalized linear model (GLM) is a flexible generalization of ordinary linear regression. Logistic regression and generalized linear model are regression models.
See Logistic regression and Generalized linear model
Goodness of fit
The goodness of fit of a statistical model describes how well it fits a set of observations.
See Logistic regression and Goodness of fit
Gradient descent
Gradient descent is a method for unconstrained mathematical optimization.
See Logistic regression and Gradient descent
Heavy-tailed distribution
In probability theory, heavy-tailed distributions are probability distributions whose tails are not exponentially bounded: that is, they have heavier tails than the exponential distribution.
See Logistic regression and Heavy-tailed distribution
Hosmer–Lemeshow test
The Hosmer–Lemeshow test is a statistical test for goodness of fit and calibration for logistic regression models.
See Logistic regression and Hosmer–Lemeshow test
Identifiability
In statistics, identifiability is a property which a model must satisfy for precise inference to be possible.
See Logistic regression and Identifiability
Independence of irrelevant alternatives
Independence of irrelevant alternatives (IIA), also known as binary independence, the independence axiom, is an axiom of decision theory and economics describing a necessary condition for rational behavior.
See Logistic regression and Independence of irrelevant alternatives
Independent and identically distributed random variables
In probability theory and statistics, a collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent.
See Logistic regression and Independent and identically distributed random variables
Indicator function
In mathematics, an indicator function or a characteristic function of a subset of a set is a function that maps elements of the subset to one, and all other elements to zero.
See Logistic regression and Indicator function
Information content
In information theory, the information content, self-information, surprisal, or Shannon information is a basic quantity derived from the probability of a particular event occurring from a random variable.
See Logistic regression and Information content
Iteratively reweighted least squares
The method of iteratively reweighted least squares (IRLS) is used to solve certain optimization problems with objective functions of the form of a ''p''-norm: \mathop_ \sum_^n \big| y_i - f_i (\boldsymbol\beta) \big|^p, by an iterative method in which each step involves solving a weighted least squares problem of the form:C.
See Logistic regression and Iteratively reweighted least squares
Jane Worcester
Jane Worcester (died 8 October 1989) was a biostatistician and epidemiologist who became the second tenured female professor, after Martha May Eliot, and the first female chair of biostatistics in the Harvard School of Public Health.
See Logistic regression and Jane Worcester
Jarrow–Turnbull model
The Jarrow–Turnbull model is a widely used "reduced-form" credit risk model.
See Logistic regression and Jarrow–Turnbull model
John Gaddum
Sir John Henry Gaddum (31 March 1900 – 30 June 1965) was an English pharmacologist who, along with Ulf von Euler, co-discovered the neuropeptide Substance P in 1931.
See Logistic regression and John Gaddum
Joseph Berkson
Joseph Berkson (14 May 1899 – 12 September 1982) was trained as a physicist (BSc 1920, College of City of New York, M.A., 1922, Columbia), physician (M.D., 1927, Johns Hopkins), and statistician (Dr.Sc., 1928, Johns Hopkins).
See Logistic regression and Joseph Berkson
Journal of Clinical Epidemiology
The Journal of Clinical Epidemiology is a peer-reviewed journal of epidemiology.
See Logistic regression and Journal of Clinical Epidemiology
Just another Gibbs sampler
Just another Gibbs sampler (JAGS) is a program for simulation from Bayesian hierarchical models using Markov chain Monte Carlo (MCMC), developed by Martyn Plummer.
See Logistic regression and Just another Gibbs sampler
Kullback–Leibler divergence
In mathematical statistics, the Kullback–Leibler (KL) divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how one probability distribution is different from a second, reference probability distribution.
See Logistic regression and Kullback–Leibler divergence
L. Gustave du Pasquier
Louis-Gustave du Pasquier (18 August 1876, Auvernier – 31 January 1957, Cornaux) was a Swiss mathematician and historian of mathematics and mathematical sciences.
See Logistic regression and L. Gustave du Pasquier
Lagrange multiplier
In mathematical optimization, the method of Lagrange multipliers is a strategy for finding the local maxima and minima of a function subject to equation constraints (i.e., subject to the condition that one or more equations have to be satisfied exactly by the chosen values of the variables).
See Logistic regression and Lagrange multiplier
Latent and observable variables
In statistics, latent variables (from Latin: present participle of lateo, “lie hidden”) are variables that can only be inferred indirectly through a mathematical model from other observable variables that can be directly observed or measured.
See Logistic regression and Latent and observable variables
Latent variable model
A latent variable model is a statistical model that relates a set of observable variables (also called manifest variables or indicators) to a set of latent variables.
See Logistic regression and Latent variable model
Likelihood function
A likelihood function (often simply called the likelihood) measures how well a statistical model explains observed data by calculating the probability of seeing that data under different parameter values of the model.
See Logistic regression and Likelihood function
Likelihood-ratio test
In statistics, the likelihood-ratio test is a hypothesis test that involves comparing the goodness of fit of two competing statistical models, typically one found by maximization over the entire parameter space and another found after imposing some constraint, based on the ratio of their likelihoods.
See Logistic regression and Likelihood-ratio test
Limited dependent variable
A limited dependent variable is a variable whose range of possible values is "restricted in some important way." In econometrics, the term is often used when estimation of the relationship between the limited dependent variable of interest and other variables requires methods that take this restriction into account.
See Logistic regression and Limited dependent variable
Limited-memory BFGS
Limited-memory BFGS (L-BFGS or LM-BFGS) is an optimization algorithm in the family of quasi-Newton methods that approximates the Broyden–Fletcher–Goldfarb–Shanno algorithm (BFGS) using a limited amount of computer memory.
See Logistic regression and Limited-memory BFGS
Linear combination
In mathematics, a linear combination is an expression constructed from a set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of x and y would be any expression of the form ax + by, where a and b are constants).
See Logistic regression and Linear combination
Linear discriminant analysis
Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features that characterizes or separates two or more classes of objects or events.
See Logistic regression and Linear discriminant analysis
Linear function (calculus)
In calculus and related areas of mathematics, a linear function from the real numbers to the real numbers is a function whose graph (in Cartesian coordinates) is a non-vertical line in the plane.
See Logistic regression and Linear function (calculus)
Linear least squares
Linear least squares (LLS) is the least squares approximation of linear functions to data.
See Logistic regression and Linear least squares
Linear model
In statistics, the term linear model refers to any model which assumes linearity in the system. Logistic regression and linear model are regression models.
See Logistic regression and Linear model
Linear predictor function
In statistics and in machine learning, a linear predictor function is a linear function (linear combination) of a set of coefficients and explanatory variables (independent variables), whose value is used to predict the outcome of a dependent variable.
See Logistic regression and Linear predictor function
Linear regression
In statistics, linear regression is a statistical model which estimates the linear relationship between a scalar response and one or more explanatory variables (also known as dependent and independent variables).
See Logistic regression and Linear regression
Local case-control sampling
In machine learning, local case-control sampling is an algorithm used to reduce the complexity of training a logistic regression classifier.
See Logistic regression and Local case-control sampling
Location parameter
In statistics, a location parameter of a probability distribution is a scalar- or vector-valued parameter x_0, which determines the "location" or shift of the distribution.
See Logistic regression and Location parameter
Logarithm
In mathematics, the logarithm is the inverse function to exponentiation.
See Logistic regression and Logarithm
Logistic distribution
In probability theory and statistics, the logistic distribution is a continuous probability distribution.
See Logistic regression and Logistic distribution
Logistic function
A logistic function or logistic curve is a common S-shaped curve (sigmoid curve) with the equation where The logistic function has domain the real numbers, the limit as x \to -\infty is 0, and the limit as x \to +\infty is L. The standard logistic function, depicted at right, where L.
See Logistic regression and Logistic function
Logistic model tree
In computer science, a logistic model tree (LMT) is a classification model with an associated supervised training algorithm that combines logistic regression (LR) and decision tree learning.
See Logistic regression and Logistic model tree
Logit
In statistics, the logit function is the quantile function associated with the standard logistic distribution.
See Logistic regression and Logit
Loss function
In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost" associated with the event.
See Logistic regression and Loss function
Loss functions for classification
In machine learning and mathematical optimization, loss functions for classification are computationally feasible loss functions representing the price paid for inaccuracy of predictions in classification problems (problems of identifying which category a particular observation belongs to).
See Logistic regression and Loss functions for classification
Lowell Reed
Lowell Jacob Reed (January 8, 1886 – April 29, 1966) was 7th president of the Johns Hopkins University in Baltimore, Maryland.
See Logistic regression and Lowell Reed
Luce's choice axiom
In probability theory, Luce's choice axiom, formulated by R. Duncan Luce (1959), states that the relative odds of selecting one item over another from a pool of many items is not affected by the presence or absence of other items in the pool.
See Logistic regression and Luce's choice axiom
Machine learning
Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of statistical algorithms that can learn from data and generalize to unseen data and thus perform tasks without explicit instructions.
See Logistic regression and Machine learning
Mark Thoma
Mark Allen Thoma (born December 15, 1956) is a macroeconomist and econometrician and a professor of economics at the Department of Economics of the University of Oregon.
See Logistic regression and Mark Thoma
Marketing
Marketing is the act of satisfying and retaining customers.
See Logistic regression and Marketing
Matching (statistics)
Matching is a statistical technique that evaluates the effect of a treatment by comparing the treated and the non-treated units in an observational study or quasi-experiment (i.e. when the treatment is not randomly assigned).
See Logistic regression and Matching (statistics)
Maximum a posteriori estimation
In Bayesian statistics, a maximum a posteriori probability (MAP) estimate is an estimate of an unknown quantity, that equals the mode of the posterior distribution.
See Logistic regression and Maximum a posteriori estimation
Maximum entropy probability distribution
In statistics and information theory, a maximum entropy probability distribution has entropy that is at least as great as that of all other members of a specified class of probability distributions.
See Logistic regression and Maximum entropy probability distribution
Maximum likelihood estimation
In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data.
See Logistic regression and Maximum likelihood estimation
Mixed logit
Mixed logit is a fully general statistical model for examining discrete choices.
See Logistic regression and Mixed logit
Mlpack
mlpack is a machine learning software library for C++, built on top of the Armadillo library and the numerical optimization library.
See Logistic regression and Mlpack
Mortgage
A mortgage loan or simply mortgage, in civil law jurisdictions known also as a hypothec loan, is a loan used either by purchasers of real property to raise funds to buy real estate, or by existing property owners to raise funds for any purpose while putting a lien on the property being mortgaged.
See Logistic regression and Mortgage
Multicollinearity
In statistics, multicollinearity or collinearity is a situation where the predictors in a regression model are linearly dependent.
See Logistic regression and Multicollinearity
Multinomial logistic regression
In statistics, multinomial logistic regression is a classification method that generalizes logistic regression to multiclass problems, i.e. with more than two possible discrete outcomes. Logistic regression and multinomial logistic regression are regression models.
See Logistic regression and Multinomial logistic regression
Myocardial infarction
A myocardial infarction (MI), commonly known as a heart attack, occurs when blood flow decreases or stops in one of the coronary arteries of the heart, causing infarction (tissue death) to the heart muscle.
See Logistic regression and Myocardial infarction
Natural language processing
Natural language processing (NLP) is an interdisciplinary subfield of computer science and artificial intelligence.
See Logistic regression and Natural language processing
Natural logarithm
The natural logarithm of a number is its logarithm to the base of the mathematical constant e, which is an irrational and transcendental number approximately equal to.
See Logistic regression and Natural logarithm
Neural network (machine learning)
In machine learning, a neural network (also artificial neural network or neural net, abbreviated ANN or NN) is a model inspired by the structure and function of biological neural networks in animal brains.
See Logistic regression and Neural network (machine learning)
Newton's method
In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a real-valued function.
See Logistic regression and Newton's method
Normal distribution
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable.
See Logistic regression and Normal distribution
Normalizing constant
In probability theory, a normalizing constant or normalizing factor is used to reduce any probability function to a probability density function with total probability of one.
See Logistic regression and Normalizing constant
Null hypothesis
In scientific research, the null hypothesis (often denoted H0) is the claim that the effect being studied does not exist.
See Logistic regression and Null hypothesis
Null model
In mathematics, for example in the study of statistical properties of graphs, a null model is a type of random object that matches one specific object in some of its features, or more generally satisfies a collection of constraints, but which is otherwise taken to be an unbiasedly random structure.
See Logistic regression and Null model
Observational study
In fields such as epidemiology, social sciences, psychology and statistics, an observational study draws inferences from a sample to a population where the independent variable is not under the control of the researcher because of ethical concerns or logistical constraints.
See Logistic regression and Observational study
Odds
In probability theory, odds provide a measure of the probability of a particular outcome.
See Logistic regression and Odds
Odds ratio
An odds ratio (OR) is a statistic that quantifies the strength of the association between two events, A and B. The odds ratio is defined as the ratio of the odds of event A taking place in the presence of B, the and odds of A in the absence of B. Due to symmetry, odds ratio reciprocally calculates the ratio of the odds of B occurring in the presence of A, and the odds of B in the absence of A.
See Logistic regression and Odds ratio
One in ten rule
In statistics, the one in ten rule is a rule of thumb for how many predictor parameters can be estimated from data when doing regression analysis (in particular proportional hazards models in survival analysis and logistic regression) while keeping the risk of overfitting and finding spurious correlations low.
See Logistic regression and One in ten rule
OpenBUGS
OpenBUGS is a software application for the Bayesian analysis of complex statistical models using Markov chain Monte Carlo (MCMC) methods.
See Logistic regression and OpenBUGS
Ordered logit
In statistics, the ordered logit model (also ordered logistic regression or proportional odds model) is an ordinal regression model—that is, a regression model for ordinal dependent variables—first considered by Peter McCullagh.
See Logistic regression and Ordered logit
Ordinary least squares
In statistics, ordinary least squares (OLS) is a type of linear least squares method for choosing the unknown parameters in a linear regression model (with fixed level-one effects of a linear function of a set of explanatory variables) by the principle of least squares: minimizing the sum of the squares of the differences between the observed dependent variable (values of the variable being observed) in the input dataset and the output of the (linear) function of the independent variable.
See Logistic regression and Ordinary least squares
Overfitting
In mathematical modeling, overfitting is "the production of an analysis that corresponds too closely or exactly to a particular set of data, and may therefore fail to fit to additional data or predict future observations reliably".
See Logistic regression and Overfitting
Parti Québécois
The paren,; PQ) is a sovereignist and social democratic provincial political party in Quebec, Canada. The PQ advocates national sovereignty for Quebec involving independence of the province of Quebec from Canada and establishing a sovereign state. The PQ has also promoted the possibility of maintaining a loose political and economic sovereignty-association between Quebec and Canada.
See Logistic regression and Parti Québécois
Partition of sums of squares
The partition of sums of squares is a concept that permeates much of inferential statistics and descriptive statistics.
See Logistic regression and Partition of sums of squares
Perceptron
In machine learning, the perceptron (or McCulloch–Pitts neuron) is an algorithm for supervised learning of binary classifiers.
See Logistic regression and Perceptron
Philosophical Transactions of the Royal Society A
Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences is a fortnightly peer-reviewed scientific journal published by the Royal Society.
See Logistic regression and Philosophical Transactions of the Royal Society A
Pierre François Verhulst
Pierre François Verhulst (28 October 1804, in Brussels – 15 February 1849, in Brussels) was a Belgian mathematician and a doctor in number theory from the University of Ghent in 1825.
See Logistic regression and Pierre François Verhulst
Poisson regression
In statistics, Poisson regression is a generalized linear model form of regression analysis used to model count data and contingency tables.
See Logistic regression and Poisson regression
Political science
Political science is the scientific study of politics.
See Logistic regression and Political science
Polynomial regression
In statistics, polynomial regression is a form of regression analysis in which the relationship between the independent variable x and the dependent variable y is modeled as an nth degree polynomial in x. Polynomial regression fits a nonlinear relationship between the value of x and the corresponding conditional mean of y, denoted E(y |x).
See Logistic regression and Polynomial regression
Population growth
Population growth is the increase in the number of people in a population or dispersed group.
See Logistic regression and Population growth
Posterior probability
The posterior probability is a type of conditional probability that results from updating the prior probability with information summarized by the likelihood via an application of Bayes' rule.
See Logistic regression and Posterior probability
Prior probability
A prior probability distribution of an uncertain quantity, often simply called the prior, is its assumed probability distribution before some evidence is taken into account.
See Logistic regression and Prior probability
Probabilistic programming
Probabilistic programming (PP) is a programming paradigm in which probabilistic models are specified and inference for these models is performed automatically.
See Logistic regression and Probabilistic programming
Probability
Probability is the branch of mathematics concerning events and numerical descriptions of how likely they are to occur.
See Logistic regression and Probability
Probability distribution
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of possible outcomes for an experiment.
See Logistic regression and Probability distribution
Probability mass function
In probability and statistics, a probability mass function (sometimes called probability function or frequency function) is a function that gives the probability that a discrete random variable is exactly equal to some value.
See Logistic regression and Probability mass function
Probit
In probability theory and statistics, the probit function is the quantile function associated with the standard normal distribution.
See Logistic regression and Probit
Probit model
In statistics, a probit model is a type of regression where the dependent variable can take only two values, for example married or not married.
See Logistic regression and Probit model
Proceedings of the National Academy of Sciences of the United States of America
Proceedings of the National Academy of Sciences of the United States of America (often abbreviated PNAS or PNAS USA) is a peer-reviewed multidisciplinary scientific journal.
PyMC
PyMC (formerly known as PyMC3) is a probabilistic programming language written in Python.
See Logistic regression and PyMC
Quasi-Newton method
Quasi-Newton methods are methods used to find either zeroes or local maxima and minima of functions, as an alternative to Newton's method.
See Logistic regression and Quasi-Newton method
Quebec
QuebecAccording to the Canadian government, Québec (with the acute accent) is the official name in Canadian French and Quebec (without the accent) is the province's official name in Canadian English is one of the thirteen provinces and territories of Canada.
See Logistic regression and Quebec
Random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events.
See Logistic regression and Random variable
Rational choice theory
Rational choice theory refers to a set of guidelines that help understand economic and social behaviour.
See Logistic regression and Rational choice theory
Raymond Pearl
Raymond Pearl (June 3, 1879 – November 17, 1940) was an American biologist, regarded as one of the founders of biogerontology.
See Logistic regression and Raymond Pearl
Real number
In mathematics, a real number is a number that can be used to measure a continuous one-dimensional quantity such as a distance, duration or temperature.
See Logistic regression and Real number
Regression analysis
In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships between a dependent variable (often called the 'outcome' or 'response' variable, or a 'label' in machine learning parlance) and one or more independent variables (often called 'predictors', 'covariates', 'explanatory variables' or 'features').
See Logistic regression and Regression analysis
Regularization (mathematics)
In mathematics, statistics, finance, and computer science, particularly in machine learning and inverse problems, regularization is a process that changes the result answer to be "simpler".
See Logistic regression and Regularization (mathematics)
Ridge regression
Ridge regression is a method of estimating the coefficients of multiple-regression models in scenarios where the independent variables are highly correlated.
See Logistic regression and Ridge regression
Robust statistics
Robust statistics are statistics that maintain their properties even if the underlying distributional assumptions are incorrect.
See Logistic regression and Robust statistics
Ronald Fisher
Sir Ronald Aylmer Fisher (17 February 1890 – 29 July 1962) was a British polymath who was active as a mathematician, statistician, biologist, geneticist, and academic.
See Logistic regression and Ronald Fisher
Scalar (mathematics)
A scalar is an element of a field which is used to define a vector space.
See Logistic regression and Scalar (mathematics)
Scale parameter
In probability theory and statistics, a scale parameter is a special kind of numerical parameter of a parametric family of probability distributions.
See Logistic regression and Scale parameter
Science (journal)
Science, also widely referred to as Science Magazine, is the peer-reviewed academic journal of the American Association for the Advancement of Science (AAAS) and one of the world's top academic journals.
See Logistic regression and Science (journal)
Separation (statistics)
In statistics, separation is a phenomenon associated with models for dichotomous or categorical outcomes, including logistic and probit regression.
See Logistic regression and Separation (statistics)
Sigmoid function
A sigmoid function is any mathematical function whose graph has a characteristic S-shaped or sigmoid curve.
See Logistic regression and Sigmoid function
Softmax function
The softmax function, also known as softargmax or normalized exponential function, converts a vector of real numbers into a probability distribution of possible outcomes.
See Logistic regression and Softmax function
Sparse matrix
In numerical analysis and scientific computing, a sparse matrix or sparse array is a matrix in which most of the elements are zero.
See Logistic regression and Sparse matrix
Spline (mathematics)
In mathematics, a spline is a function defined piecewise by polynomials.
See Logistic regression and Spline (mathematics)
Stan (software)
Stan is a probabilistic programming language for statistical inference written in C++.
See Logistic regression and Stan (software)
Statistical classification
When classification is performed by a computer, statistical methods are normally used to develop the algorithm.
See Logistic regression and Statistical classification
Statistical data type
In statistics, groups of individual data points may be classified as belonging to any of various statistical data types, e.g. categorical ("red", "blue", "green"), real number, odd number (1,3,5) etc.
See Logistic regression and Statistical data type
Statistical model
A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of sample data (and similar data from a larger population).
See Logistic regression and Statistical model
Statistics
Statistics (from German: Statistik, "description of a state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data.
See Logistic regression and Statistics
Step function
In mathematics, a function on the real numbers is called a step function if it can be written as a finite linear combination of indicator functions of intervals.
See Logistic regression and Step function
Stratification (clinical trials)
Stratification of clinical trials is the partitioning of subjects and results by a factor other than the treatment given.
See Logistic regression and Stratification (clinical trials)
Supervised learning
Supervised learning (SL) is a paradigm in machine learning where input objects (for example, a vector of predictor variables) and a desired output value (also known as human-labeled supervisory signal) train a model.
See Logistic regression and Supervised learning
Trauma Quality Improvement Program
The Trauma Quality Improvement Program (TQIP) was initiated in 2008 by the American College of Surgeons Committee on Trauma.
See Logistic regression and Trauma Quality Improvement Program
Type I and type II errors
In statistical hypothesis testing, a type I error, or a false positive, is the rejection of the null hypothesis when it is actually true.
See Logistic regression and Type I and type II errors
Udny Yule
George Udny Yule, CBE, FRS (18 February 1871 – 26 June 1951), usually known as Udny Yule, was a British statistician, particularly known for the Yule distribution and proposing the preferential attachment model for random graphs.
See Logistic regression and Udny Yule
Unit of measurement
A unit of measurement, or unit of measure, is a definite magnitude of a quantity, defined and adopted by convention or by law, that is used as a standard for measurement of the same kind of quantity.
See Logistic regression and Unit of measurement
Utility
In economics, utility is a measure of the satisfaction that a certain person has from a certain state of the world.
See Logistic regression and Utility
Value (mathematics)
In mathematics, value may refer to several, strongly related notions.
See Logistic regression and Value (mathematics)
Variational Bayesian methods
Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning.
See Logistic regression and Variational Bayesian methods
Wald test
In statistics, the Wald test (named after Abraham Wald) assesses constraints on statistical parameters based on the weighted distance between the unrestricted estimate and its hypothesized value under the null hypothesis, where the weight is the precision of the estimate.
See Logistic regression and Wald test
Wilhelm Ostwald
Friedrich Wilhelm Ostwald (4 April 1932) was a Baltic German chemist and philosopher.
See Logistic regression and Wilhelm Ostwald
Y-intercept
In analytic geometry, using the common convention that the horizontal axis represents a variable x and the vertical axis represents a variable y, a y-intercept or vertical intercept is a point where the graph of a function or relation intersects the y-axis of the coordinate system.
See Logistic regression and Y-intercept
See also
Predictive analytics
- Cambridge Analytica
- Civis Analytics
- Convergent cross mapping
- Empirical dynamic modeling
- Geolitica
- Logistic regression
- Metaculus
- Predictive analytics
- Predictive mean matching
- Predictive modelling
Regression models
- Additive model
- Bradley–Terry model
- Censored regression model
- Choice modelling
- Conditional change model
- Discriminative model
- EM algorithm and GMM model
- Errors-in-variables models
- Factor regression model
- Fay–Herriot model
- First-hitting-time model
- Fixed effects model
- GHK algorithm
- General linear model
- Generalized additive model
- Generalized linear array model
- Generalized linear model
- Generalized linear models
- Hedonic regression
- Hierarchical generalized linear model
- History index model
- Linear model
- Logistic regression
- Marginal model
- Mixed model
- Multi-attribute global inference of quality
- Multilevel model
- Multilevel modeling for repeated measures
- Multilevel regression with poststratification
- Multinomial logistic regression
- Multivariate probit model
- Nonlinear mixed-effects model
- Polynomial and rational function modeling
- Proper linear model
- Random effects model
- Regression dilution
- Response modeling methodology
- Segmented regression
- Simultaneous equations model
- Sinusoidal model
- Stock sampling
- Structural equation modeling
- Threshold model
- Time series models
- Tobit model
- Total least squares
- Truncated normal hurdle model
- Truncated regression model
- Vector generalized linear model
References
[1] https://en.wikipedia.org/wiki/Logistic_regression
Also known as Applications of logistic regression, Binary logit model, Conditional logit analysis, Logit model, Logit regression.
, Edwin Bidwell Wilson, Emergency management, Engineering, Entropy (information theory), Error function, Estimation theory, Euler numbers, Expectation propagation, Expected value, Exponential family, Exponential function, Generalized linear model, Goodness of fit, Gradient descent, Heavy-tailed distribution, Hosmer–Lemeshow test, Identifiability, Independence of irrelevant alternatives, Independent and identically distributed random variables, Indicator function, Information content, Iteratively reweighted least squares, Jane Worcester, Jarrow–Turnbull model, John Gaddum, Joseph Berkson, Journal of Clinical Epidemiology, Just another Gibbs sampler, Kullback–Leibler divergence, L. Gustave du Pasquier, Lagrange multiplier, Latent and observable variables, Latent variable model, Likelihood function, Likelihood-ratio test, Limited dependent variable, Limited-memory BFGS, Linear combination, Linear discriminant analysis, Linear function (calculus), Linear least squares, Linear model, Linear predictor function, Linear regression, Local case-control sampling, Location parameter, Logarithm, Logistic distribution, Logistic function, Logistic model tree, Logit, Loss function, Loss functions for classification, Lowell Reed, Luce's choice axiom, Machine learning, Mark Thoma, Marketing, Matching (statistics), Maximum a posteriori estimation, Maximum entropy probability distribution, Maximum likelihood estimation, Mixed logit, Mlpack, Mortgage, Multicollinearity, Multinomial logistic regression, Myocardial infarction, Natural language processing, Natural logarithm, Neural network (machine learning), Newton's method, Normal distribution, Normalizing constant, Null hypothesis, Null model, Observational study, Odds, Odds ratio, One in ten rule, OpenBUGS, Ordered logit, Ordinary least squares, Overfitting, Parti Québécois, Partition of sums of squares, Perceptron, Philosophical Transactions of the Royal Society A, Pierre François Verhulst, Poisson regression, Political science, Polynomial regression, Population growth, Posterior probability, Prior probability, Probabilistic programming, Probability, Probability distribution, Probability mass function, Probit, Probit model, Proceedings of the National Academy of Sciences of the United States of America, PyMC, Quasi-Newton method, Quebec, Random variable, Rational choice theory, Raymond Pearl, Real number, Regression analysis, Regularization (mathematics), Ridge regression, Robust statistics, Ronald Fisher, Scalar (mathematics), Scale parameter, Science (journal), Separation (statistics), Sigmoid function, Softmax function, Sparse matrix, Spline (mathematics), Stan (software), Statistical classification, Statistical data type, Statistical model, Statistics, Step function, Stratification (clinical trials), Supervised learning, Trauma Quality Improvement Program, Type I and type II errors, Udny Yule, Unit of measurement, Utility, Value (mathematics), Variational Bayesian methods, Wald test, Wilhelm Ostwald, Y-intercept.