en.unionpedia.org

Logistic regression, the Glossary

Index Logistic regression

In statistics, the logistic model (or logit model) is a statistical model that models the log-odds of an event as a linear combination of one or more independent variables.[1]

Table of Contents

  1. 190 relations: Adolphe Quetelet, Analytic function, Autocatalysis, Backpropagation, Bayesian statistics, Bernoulli distribution, Bernoulli trial, Binary classification, Binary data, Binary regression, Binomial distribution, Bioassay, Blood test, Body mass index, Brier score, Built environment, C (programming language), C++, Cambridge University Press, Canada, Canonical form, Cardinal number, Catalysis, Categorical variable, Chester Ittner Bliss, Chi-squared test, Coefficient of determination, Computer science, Conditional entropy, Conditional logistic regression, Conditional random field, Conjugate prior, Contingency table, Continuous or discrete variable, Coronary artery disease, Cross-entropy, Cumulative distribution function, Daniel McFadden, David Cox (statistician), Degrees of freedom (statistics), Dependent and independent variables, Design matrix, Deviance (statistics), Diabetes, Discrete choice, Dot product, Dummy variable (statistics), E (mathematical constant), Econometrics, Economics, ... Expand index (140 more) »

  2. Predictive analytics
  3. Regression models

Adolphe Quetelet

Lambert Adolphe Jacques Quetelet FRSF or FRSE (22 February 1796 – 17 February 1874) was a Belgian astronomer, mathematician, statistician and sociologist who founded and directed the Brussels Observatory and was influential in introducing statistical methods to the social sciences.

See Logistic regression and Adolphe Quetelet

Analytic function

In mathematics, an analytic function is a function that is locally given by a convergent power series.

See Logistic regression and Analytic function

Autocatalysis

In chemistry, a chemical reaction is said to be autocatalytic if one of the reaction products is also a catalyst for the same reaction.

See Logistic regression and Autocatalysis

Backpropagation

In machine learning, backpropagation is a gradient estimation method used to train neural network models.

See Logistic regression and Backpropagation

Bayesian statistics

Bayesian statistics is a theory in the field of statistics based on the Bayesian interpretation of probability, where probability expresses a degree of belief in an event.

See Logistic regression and Bayesian statistics

Bernoulli distribution

In probability theory and statistics, the Bernoulli distribution, named after Swiss mathematician Jacob Bernoulli, is the discrete probability distribution of a random variable which takes the value 1 with probability p and the value 0 with probability q.

See Logistic regression and Bernoulli distribution

Bernoulli trial

In the theory of probability and statistics, a Bernoulli trial (or binomial trial) is a random experiment with exactly two possible outcomes, "success" and "failure", in which the probability of success is the same every time the experiment is conducted.

See Logistic regression and Bernoulli trial

Binary classification

Binary classification is the task of classifying the elements of a set into one of two groups (each called class).

See Logistic regression and Binary classification

Binary data

Binary data is data whose unit can take on only two possible states.

See Logistic regression and Binary data

Binary regression

In statistics, specifically regression analysis, a binary regression estimates a relationship between one or more explanatory variables and a single output binary variable.

See Logistic regression and Binary regression

Binomial distribution

In probability theory and statistics, the binomial distribution with parameters and is the discrete probability distribution of the number of successes in a sequence of independent experiments, each asking a yes–no question, and each with its own Boolean-valued outcome: success (with probability) or failure (with probability).

See Logistic regression and Binomial distribution

Bioassay

A bioassay is an analytical method to determine the potency or effect of a substance by its effect on living animals or plants (in vivo), or on living cells or tissues (in vitro).

See Logistic regression and Bioassay

Blood test

A blood test is a laboratory analysis performed on a blood sample that is usually extracted from a vein in the arm using a hypodermic needle, or via fingerprick.

See Logistic regression and Blood test

Body mass index

Body mass index (BMI) is a value derived from the mass (weight) and height of a person.

See Logistic regression and Body mass index

Brier score

The Brier Score is a ''strictly proper score function'' or ''strictly proper scoring rule'' that measures the accuracy of probabilistic predictions.

See Logistic regression and Brier score

Built environment

The term built environment refers to human-made conditions and is often used in architecture, landscape architecture, urban planning, public health, sociology, and anthropology, among others.

See Logistic regression and Built environment

C (programming language)

C (pronounced – like the letter c) is a general-purpose programming language.

See Logistic regression and C (programming language)

C++

C++ (pronounced "C plus plus" and sometimes abbreviated as CPP) is a high-level, general-purpose programming language created by Danish computer scientist Bjarne Stroustrup.

See Logistic regression and C++

Cambridge University Press

Cambridge University Press is the university press of the University of Cambridge.

See Logistic regression and Cambridge University Press

Canada

Canada is a country in North America.

See Logistic regression and Canada

Canonical form

In mathematics and computer science, a canonical, normal, or standard form of a mathematical object is a standard way of presenting that object as a mathematical expression.

See Logistic regression and Canonical form

Cardinal number

In mathematics, a cardinal number, or cardinal for short, is what is commonly called the number of elements of a set.

See Logistic regression and Cardinal number

Catalysis

Catalysis is the increase in rate of a chemical reaction due to an added substance known as a catalyst.

See Logistic regression and Catalysis

Categorical variable

In statistics, a categorical variable (also called qualitative variable) is a variable that can take on one of a limited, and usually fixed, number of possible values, assigning each individual or other unit of observation to a particular group or nominal category on the basis of some qualitative property.

See Logistic regression and Categorical variable

Chester Ittner Bliss

Chester Ittner Bliss (February 1, 1899 – March 14, 1979) was primarily a biologist, who is best known for his contributions to statistics.

See Logistic regression and Chester Ittner Bliss

Chi-squared test

A chi-squared test (also chi-square or test) is a statistical hypothesis test used in the analysis of contingency tables when the sample sizes are large.

See Logistic regression and Chi-squared test

Coefficient of determination

In statistics, the coefficient of determination, denoted R2 or r2 and pronounced "R squared", is the proportion of the variation in the dependent variable that is predictable from the independent variable(s).

See Logistic regression and Coefficient of determination

Computer science

Computer science is the study of computation, information, and automation.

See Logistic regression and Computer science

Conditional entropy

In information theory, the conditional entropy quantifies the amount of information needed to describe the outcome of a random variable Y given that the value of another random variable X is known.

See Logistic regression and Conditional entropy

Conditional logistic regression

Conditional logistic regression is an extension of logistic regression that allows one to account for stratification and matching.

See Logistic regression and Conditional logistic regression

Conditional random field

Conditional random fields (CRFs) are a class of statistical modeling methods often applied in pattern recognition and machine learning and used for structured prediction.

See Logistic regression and Conditional random field

Conjugate prior

In Bayesian probability theory, if, given a likelihood function p(x \mid \theta), the posterior distribution p(\theta \mid x) is in the same probability distribution family as the prior probability distribution p(\theta), the prior and posterior are then called conjugate distributions with respect to that likelihood function and the prior is called a conjugate prior for the likelihood function p(x \mid \theta).

See Logistic regression and Conjugate prior

Contingency table

In statistics, a contingency table (also known as a cross tabulation or crosstab) is a type of table in a matrix format that displays the multivariate frequency distribution of the variables.

See Logistic regression and Contingency table

Continuous or discrete variable

In mathematics and statistics, a quantitative variable may be continuous or discrete if they are typically obtained by measuring or counting, respectively.

See Logistic regression and Continuous or discrete variable

Coronary artery disease

Coronary artery disease (CAD), also called coronary heart disease (CHD), ischemic heart disease (IHD), myocardial ischemia, or simply heart disease, involves the reduction of blood flow to the cardiac muscle due to build-up of atherosclerotic plaque in the arteries of the heart.

See Logistic regression and Coronary artery disease

Cross-entropy

In information theory, the cross-entropy between two probability distributions p and q, over the same underlying set of events, measures the average number of bits needed to identify an event drawn from the set when the coding scheme used for the set is optimized for an estimated probability distribution q, rather than the true distribution p.

See Logistic regression and Cross-entropy

Cumulative distribution function

In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Every probability distribution supported on the real numbers, discrete or "mixed" as well as continuous, is uniquely identified by a right-continuous monotone increasing function (a càdlàg function) F \colon \mathbb R \rightarrow satisfying \lim_F(x).

See Logistic regression and Cumulative distribution function

Daniel McFadden

Daniel Little McFadden (born July 29, 1937) is an American econometrician who shared the 2000 Nobel Memorial Prize in Economic Sciences with James Heckman.

See Logistic regression and Daniel McFadden

David Cox (statistician)

Sir David Roxbee Cox (15 July 1924 – 18 January 2022) was a British statistician and educator.

See Logistic regression and David Cox (statistician)

Degrees of freedom (statistics)

In statistics, the number of degrees of freedom is the number of values in the final calculation of a statistic that are free to vary.

See Logistic regression and Degrees of freedom (statistics)

Dependent and independent variables

A variable is considered dependent if it depends on an independent variable.

See Logistic regression and Dependent and independent variables

Design matrix

In statistics and in particular in regression analysis, a design matrix, also known as model matrix or regressor matrix and often denoted by X, is a matrix of values of explanatory variables of a set of objects.

See Logistic regression and Design matrix

Deviance (statistics)

In statistics, deviance is a goodness-of-fit statistic for a statistical model; it is often used for statistical hypothesis testing.

See Logistic regression and Deviance (statistics)

Diabetes

Diabetes mellitus, often known simply as diabetes, is a group of common endocrine diseases characterized by sustained high blood sugar levels.

See Logistic regression and Diabetes

Discrete choice

In economics, discrete choice models, or qualitative choice models, describe, explain, and predict choices between two or more discrete alternatives, such as entering or not entering the labor market, or choosing between modes of transport.

See Logistic regression and Discrete choice

Dot product

In mathematics, the dot product or scalar productThe term scalar product means literally "product with a scalar as a result".

See Logistic regression and Dot product

Dummy variable (statistics)

In regression analysis, a dummy variable (also known as indicator variable or just dummy) is one that takes a binary value (0 or 1) to indicate the absence or presence of some categorical effect that may be expected to shift the outcome.

See Logistic regression and Dummy variable (statistics)

E (mathematical constant)

The number is a mathematical constant approximately equal to 2.71828 that can be characterized in many ways.

See Logistic regression and E (mathematical constant)

Econometrics

Econometrics is an application of statistical methods to economic data in order to give empirical content to economic relationships.

See Logistic regression and Econometrics

Economics

Economics is a social science that studies the production, distribution, and consumption of goods and services.

See Logistic regression and Economics

Edwin Bidwell Wilson

Edwin Bidwell Wilson (April 25, 1879 – December 28, 1964) was an American mathematician, statistician, physicist and general polymath.

See Logistic regression and Edwin Bidwell Wilson

Emergency management

Emergency management (also disaster management) is a science and a system charged with creating the framework within which communities reduce vulnerability to hazards and cope with disasters.

See Logistic regression and Emergency management

Engineering

Engineering is the practice of using natural science, mathematics, and the engineering design process to solve technical problems, increase efficiency and productivity, and improve systems.

See Logistic regression and Engineering

Entropy (information theory)

In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes.

See Logistic regression and Entropy (information theory)

Error function

In mathematics, the error function (also called the Gauss error function), often denoted by, is a function defined as: \operatorname z.

See Logistic regression and Error function

Estimation theory

Estimation theory is a branch of statistics that deals with estimating the values of parameters based on measured empirical data that has a random component.

See Logistic regression and Estimation theory

Euler numbers

In mathematics, the Euler numbers are a sequence En of integers defined by the Taylor series expansion where \cosh (t) is the hyperbolic cosine function.

See Logistic regression and Euler numbers

Expectation propagation

Expectation propagation (EP) is a technique in Bayesian machine learning.

See Logistic regression and Expectation propagation

Expected value

In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first moment) is a generalization of the weighted average.

See Logistic regression and Expected value

Exponential family

In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below.

See Logistic regression and Exponential family

Exponential function

The exponential function is a mathematical function denoted by f(x).

See Logistic regression and Exponential function

Generalized linear model

In statistics, a generalized linear model (GLM) is a flexible generalization of ordinary linear regression. Logistic regression and generalized linear model are regression models.

See Logistic regression and Generalized linear model

Goodness of fit

The goodness of fit of a statistical model describes how well it fits a set of observations.

See Logistic regression and Goodness of fit

Gradient descent

Gradient descent is a method for unconstrained mathematical optimization.

See Logistic regression and Gradient descent

Heavy-tailed distribution

In probability theory, heavy-tailed distributions are probability distributions whose tails are not exponentially bounded: that is, they have heavier tails than the exponential distribution.

See Logistic regression and Heavy-tailed distribution

Hosmer–Lemeshow test

The Hosmer–Lemeshow test is a statistical test for goodness of fit and calibration for logistic regression models.

See Logistic regression and Hosmer–Lemeshow test

Identifiability

In statistics, identifiability is a property which a model must satisfy for precise inference to be possible.

See Logistic regression and Identifiability

Independence of irrelevant alternatives

Independence of irrelevant alternatives (IIA), also known as binary independence, the independence axiom, is an axiom of decision theory and economics describing a necessary condition for rational behavior.

See Logistic regression and Independence of irrelevant alternatives

Independent and identically distributed random variables

In probability theory and statistics, a collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent.

See Logistic regression and Independent and identically distributed random variables

Indicator function

In mathematics, an indicator function or a characteristic function of a subset of a set is a function that maps elements of the subset to one, and all other elements to zero.

See Logistic regression and Indicator function

Information content

In information theory, the information content, self-information, surprisal, or Shannon information is a basic quantity derived from the probability of a particular event occurring from a random variable.

See Logistic regression and Information content

Iteratively reweighted least squares

The method of iteratively reweighted least squares (IRLS) is used to solve certain optimization problems with objective functions of the form of a ''p''-norm: \mathop_ \sum_^n \big| y_i - f_i (\boldsymbol\beta) \big|^p, by an iterative method in which each step involves solving a weighted least squares problem of the form:C.

See Logistic regression and Iteratively reweighted least squares

Jane Worcester

Jane Worcester (died 8 October 1989) was a biostatistician and epidemiologist who became the second tenured female professor, after Martha May Eliot, and the first female chair of biostatistics in the Harvard School of Public Health.

See Logistic regression and Jane Worcester

Jarrow–Turnbull model

The Jarrow–Turnbull model is a widely used "reduced-form" credit risk model.

See Logistic regression and Jarrow–Turnbull model

John Gaddum

Sir John Henry Gaddum (31 March 1900 – 30 June 1965) was an English pharmacologist who, along with Ulf von Euler, co-discovered the neuropeptide Substance P in 1931.

See Logistic regression and John Gaddum

Joseph Berkson

Joseph Berkson (14 May 1899 – 12 September 1982) was trained as a physicist (BSc 1920, College of City of New York, M.A., 1922, Columbia), physician (M.D., 1927, Johns Hopkins), and statistician (Dr.Sc., 1928, Johns Hopkins).

See Logistic regression and Joseph Berkson

Journal of Clinical Epidemiology

The Journal of Clinical Epidemiology is a peer-reviewed journal of epidemiology.

See Logistic regression and Journal of Clinical Epidemiology

Just another Gibbs sampler

Just another Gibbs sampler (JAGS) is a program for simulation from Bayesian hierarchical models using Markov chain Monte Carlo (MCMC), developed by Martyn Plummer.

See Logistic regression and Just another Gibbs sampler

Kullback–Leibler divergence

In mathematical statistics, the Kullback–Leibler (KL) divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how one probability distribution is different from a second, reference probability distribution.

See Logistic regression and Kullback–Leibler divergence

L. Gustave du Pasquier

Louis-Gustave du Pasquier (18 August 1876, Auvernier – 31 January 1957, Cornaux) was a Swiss mathematician and historian of mathematics and mathematical sciences.

See Logistic regression and L. Gustave du Pasquier

Lagrange multiplier

In mathematical optimization, the method of Lagrange multipliers is a strategy for finding the local maxima and minima of a function subject to equation constraints (i.e., subject to the condition that one or more equations have to be satisfied exactly by the chosen values of the variables).

See Logistic regression and Lagrange multiplier

Latent and observable variables

In statistics, latent variables (from Latin: present participle of lateo, “lie hidden”) are variables that can only be inferred indirectly through a mathematical model from other observable variables that can be directly observed or measured.

See Logistic regression and Latent and observable variables

Latent variable model

A latent variable model is a statistical model that relates a set of observable variables (also called manifest variables or indicators) to a set of latent variables.

See Logistic regression and Latent variable model

Likelihood function

A likelihood function (often simply called the likelihood) measures how well a statistical model explains observed data by calculating the probability of seeing that data under different parameter values of the model.

See Logistic regression and Likelihood function

Likelihood-ratio test

In statistics, the likelihood-ratio test is a hypothesis test that involves comparing the goodness of fit of two competing statistical models, typically one found by maximization over the entire parameter space and another found after imposing some constraint, based on the ratio of their likelihoods.

See Logistic regression and Likelihood-ratio test

Limited dependent variable

A limited dependent variable is a variable whose range of possible values is "restricted in some important way." In econometrics, the term is often used when estimation of the relationship between the limited dependent variable of interest and other variables requires methods that take this restriction into account.

See Logistic regression and Limited dependent variable

Limited-memory BFGS

Limited-memory BFGS (L-BFGS or LM-BFGS) is an optimization algorithm in the family of quasi-Newton methods that approximates the Broyden–Fletcher–Goldfarb–Shanno algorithm (BFGS) using a limited amount of computer memory.

See Logistic regression and Limited-memory BFGS

Linear combination

In mathematics, a linear combination is an expression constructed from a set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of x and y would be any expression of the form ax + by, where a and b are constants).

See Logistic regression and Linear combination

Linear discriminant analysis

Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features that characterizes or separates two or more classes of objects or events.

See Logistic regression and Linear discriminant analysis

Linear function (calculus)

In calculus and related areas of mathematics, a linear function from the real numbers to the real numbers is a function whose graph (in Cartesian coordinates) is a non-vertical line in the plane.

See Logistic regression and Linear function (calculus)

Linear least squares

Linear least squares (LLS) is the least squares approximation of linear functions to data.

See Logistic regression and Linear least squares

Linear model

In statistics, the term linear model refers to any model which assumes linearity in the system. Logistic regression and linear model are regression models.

See Logistic regression and Linear model

Linear predictor function

In statistics and in machine learning, a linear predictor function is a linear function (linear combination) of a set of coefficients and explanatory variables (independent variables), whose value is used to predict the outcome of a dependent variable.

See Logistic regression and Linear predictor function

Linear regression

In statistics, linear regression is a statistical model which estimates the linear relationship between a scalar response and one or more explanatory variables (also known as dependent and independent variables).

See Logistic regression and Linear regression

Local case-control sampling

In machine learning, local case-control sampling is an algorithm used to reduce the complexity of training a logistic regression classifier.

See Logistic regression and Local case-control sampling

Location parameter

In statistics, a location parameter of a probability distribution is a scalar- or vector-valued parameter x_0, which determines the "location" or shift of the distribution.

See Logistic regression and Location parameter

Logarithm

In mathematics, the logarithm is the inverse function to exponentiation.

See Logistic regression and Logarithm

Logistic distribution

In probability theory and statistics, the logistic distribution is a continuous probability distribution.

See Logistic regression and Logistic distribution

Logistic function

A logistic function or logistic curve is a common S-shaped curve (sigmoid curve) with the equation where The logistic function has domain the real numbers, the limit as x \to -\infty is 0, and the limit as x \to +\infty is L. The standard logistic function, depicted at right, where L.

See Logistic regression and Logistic function

Logistic model tree

In computer science, a logistic model tree (LMT) is a classification model with an associated supervised training algorithm that combines logistic regression (LR) and decision tree learning.

See Logistic regression and Logistic model tree

Logit

In statistics, the logit function is the quantile function associated with the standard logistic distribution.

See Logistic regression and Logit

Loss function

In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost" associated with the event.

See Logistic regression and Loss function

Loss functions for classification

In machine learning and mathematical optimization, loss functions for classification are computationally feasible loss functions representing the price paid for inaccuracy of predictions in classification problems (problems of identifying which category a particular observation belongs to).

See Logistic regression and Loss functions for classification

Lowell Reed

Lowell Jacob Reed (January 8, 1886 – April 29, 1966) was 7th president of the Johns Hopkins University in Baltimore, Maryland.

See Logistic regression and Lowell Reed

Luce's choice axiom

In probability theory, Luce's choice axiom, formulated by R. Duncan Luce (1959), states that the relative odds of selecting one item over another from a pool of many items is not affected by the presence or absence of other items in the pool.

See Logistic regression and Luce's choice axiom

Machine learning

Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of statistical algorithms that can learn from data and generalize to unseen data and thus perform tasks without explicit instructions.

See Logistic regression and Machine learning

Mark Thoma

Mark Allen Thoma (born December 15, 1956) is a macroeconomist and econometrician and a professor of economics at the Department of Economics of the University of Oregon.

See Logistic regression and Mark Thoma

Marketing

Marketing is the act of satisfying and retaining customers.

See Logistic regression and Marketing

Matching (statistics)

Matching is a statistical technique that evaluates the effect of a treatment by comparing the treated and the non-treated units in an observational study or quasi-experiment (i.e. when the treatment is not randomly assigned).

See Logistic regression and Matching (statistics)

Maximum a posteriori estimation

In Bayesian statistics, a maximum a posteriori probability (MAP) estimate is an estimate of an unknown quantity, that equals the mode of the posterior distribution.

See Logistic regression and Maximum a posteriori estimation

Maximum entropy probability distribution

In statistics and information theory, a maximum entropy probability distribution has entropy that is at least as great as that of all other members of a specified class of probability distributions.

See Logistic regression and Maximum entropy probability distribution

Maximum likelihood estimation

In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data.

See Logistic regression and Maximum likelihood estimation

Mixed logit

Mixed logit is a fully general statistical model for examining discrete choices.

See Logistic regression and Mixed logit

Mlpack

mlpack is a machine learning software library for C++, built on top of the Armadillo library and the numerical optimization library.

See Logistic regression and Mlpack

Mortgage

A mortgage loan or simply mortgage, in civil law jurisdictions known also as a hypothec loan, is a loan used either by purchasers of real property to raise funds to buy real estate, or by existing property owners to raise funds for any purpose while putting a lien on the property being mortgaged.

See Logistic regression and Mortgage

Multicollinearity

In statistics, multicollinearity or collinearity is a situation where the predictors in a regression model are linearly dependent.

See Logistic regression and Multicollinearity

Multinomial logistic regression

In statistics, multinomial logistic regression is a classification method that generalizes logistic regression to multiclass problems, i.e. with more than two possible discrete outcomes. Logistic regression and multinomial logistic regression are regression models.

See Logistic regression and Multinomial logistic regression

Myocardial infarction

A myocardial infarction (MI), commonly known as a heart attack, occurs when blood flow decreases or stops in one of the coronary arteries of the heart, causing infarction (tissue death) to the heart muscle.

See Logistic regression and Myocardial infarction

Natural language processing

Natural language processing (NLP) is an interdisciplinary subfield of computer science and artificial intelligence.

See Logistic regression and Natural language processing

Natural logarithm

The natural logarithm of a number is its logarithm to the base of the mathematical constant e, which is an irrational and transcendental number approximately equal to.

See Logistic regression and Natural logarithm

Neural network (machine learning)

In machine learning, a neural network (also artificial neural network or neural net, abbreviated ANN or NN) is a model inspired by the structure and function of biological neural networks in animal brains.

See Logistic regression and Neural network (machine learning)

Newton's method

In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a real-valued function.

See Logistic regression and Newton's method

Normal distribution

In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable.

See Logistic regression and Normal distribution

Normalizing constant

In probability theory, a normalizing constant or normalizing factor is used to reduce any probability function to a probability density function with total probability of one.

See Logistic regression and Normalizing constant

Null hypothesis

In scientific research, the null hypothesis (often denoted H0) is the claim that the effect being studied does not exist.

See Logistic regression and Null hypothesis

Null model

In mathematics, for example in the study of statistical properties of graphs, a null model is a type of random object that matches one specific object in some of its features, or more generally satisfies a collection of constraints, but which is otherwise taken to be an unbiasedly random structure.

See Logistic regression and Null model

Observational study

In fields such as epidemiology, social sciences, psychology and statistics, an observational study draws inferences from a sample to a population where the independent variable is not under the control of the researcher because of ethical concerns or logistical constraints.

See Logistic regression and Observational study

Odds

In probability theory, odds provide a measure of the probability of a particular outcome.

See Logistic regression and Odds

Odds ratio

An odds ratio (OR) is a statistic that quantifies the strength of the association between two events, A and B. The odds ratio is defined as the ratio of the odds of event A taking place in the presence of B, the and odds of A in the absence of B. Due to symmetry, odds ratio reciprocally calculates the ratio of the odds of B occurring in the presence of A, and the odds of B in the absence of A.

See Logistic regression and Odds ratio

One in ten rule

In statistics, the one in ten rule is a rule of thumb for how many predictor parameters can be estimated from data when doing regression analysis (in particular proportional hazards models in survival analysis and logistic regression) while keeping the risk of overfitting and finding spurious correlations low.

See Logistic regression and One in ten rule

OpenBUGS

OpenBUGS is a software application for the Bayesian analysis of complex statistical models using Markov chain Monte Carlo (MCMC) methods.

See Logistic regression and OpenBUGS

Ordered logit

In statistics, the ordered logit model (also ordered logistic regression or proportional odds model) is an ordinal regression model—that is, a regression model for ordinal dependent variables—first considered by Peter McCullagh.

See Logistic regression and Ordered logit

Ordinary least squares

In statistics, ordinary least squares (OLS) is a type of linear least squares method for choosing the unknown parameters in a linear regression model (with fixed level-one effects of a linear function of a set of explanatory variables) by the principle of least squares: minimizing the sum of the squares of the differences between the observed dependent variable (values of the variable being observed) in the input dataset and the output of the (linear) function of the independent variable.

See Logistic regression and Ordinary least squares

Overfitting

In mathematical modeling, overfitting is "the production of an analysis that corresponds too closely or exactly to a particular set of data, and may therefore fail to fit to additional data or predict future observations reliably".

See Logistic regression and Overfitting

Parti Québécois

The paren,; PQ) is a sovereignist and social democratic provincial political party in Quebec, Canada. The PQ advocates national sovereignty for Quebec involving independence of the province of Quebec from Canada and establishing a sovereign state. The PQ has also promoted the possibility of maintaining a loose political and economic sovereignty-association between Quebec and Canada.

See Logistic regression and Parti Québécois

Partition of sums of squares

The partition of sums of squares is a concept that permeates much of inferential statistics and descriptive statistics.

See Logistic regression and Partition of sums of squares

Perceptron

In machine learning, the perceptron (or McCulloch–Pitts neuron) is an algorithm for supervised learning of binary classifiers.

See Logistic regression and Perceptron

Philosophical Transactions of the Royal Society A

Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences is a fortnightly peer-reviewed scientific journal published by the Royal Society.

See Logistic regression and Philosophical Transactions of the Royal Society A

Pierre François Verhulst

Pierre François Verhulst (28 October 1804, in Brussels – 15 February 1849, in Brussels) was a Belgian mathematician and a doctor in number theory from the University of Ghent in 1825.

See Logistic regression and Pierre François Verhulst

Poisson regression

In statistics, Poisson regression is a generalized linear model form of regression analysis used to model count data and contingency tables.

See Logistic regression and Poisson regression

Political science

Political science is the scientific study of politics.

See Logistic regression and Political science

Polynomial regression

In statistics, polynomial regression is a form of regression analysis in which the relationship between the independent variable x and the dependent variable y is modeled as an nth degree polynomial in x. Polynomial regression fits a nonlinear relationship between the value of x and the corresponding conditional mean of y, denoted E(y |x).

See Logistic regression and Polynomial regression

Population growth

Population growth is the increase in the number of people in a population or dispersed group.

See Logistic regression and Population growth

Posterior probability

The posterior probability is a type of conditional probability that results from updating the prior probability with information summarized by the likelihood via an application of Bayes' rule.

See Logistic regression and Posterior probability

Prior probability

A prior probability distribution of an uncertain quantity, often simply called the prior, is its assumed probability distribution before some evidence is taken into account.

See Logistic regression and Prior probability

Probabilistic programming

Probabilistic programming (PP) is a programming paradigm in which probabilistic models are specified and inference for these models is performed automatically.

See Logistic regression and Probabilistic programming

Probability

Probability is the branch of mathematics concerning events and numerical descriptions of how likely they are to occur.

See Logistic regression and Probability

Probability distribution

In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of possible outcomes for an experiment.

See Logistic regression and Probability distribution

Probability mass function

In probability and statistics, a probability mass function (sometimes called probability function or frequency function) is a function that gives the probability that a discrete random variable is exactly equal to some value.

See Logistic regression and Probability mass function

Probit

In probability theory and statistics, the probit function is the quantile function associated with the standard normal distribution.

See Logistic regression and Probit

Probit model

In statistics, a probit model is a type of regression where the dependent variable can take only two values, for example married or not married.

See Logistic regression and Probit model

Proceedings of the National Academy of Sciences of the United States of America

Proceedings of the National Academy of Sciences of the United States of America (often abbreviated PNAS or PNAS USA) is a peer-reviewed multidisciplinary scientific journal.

See Logistic regression and Proceedings of the National Academy of Sciences of the United States of America

PyMC

PyMC (formerly known as PyMC3) is a probabilistic programming language written in Python.

See Logistic regression and PyMC

Quasi-Newton method

Quasi-Newton methods are methods used to find either zeroes or local maxima and minima of functions, as an alternative to Newton's method.

See Logistic regression and Quasi-Newton method

Quebec

QuebecAccording to the Canadian government, Québec (with the acute accent) is the official name in Canadian French and Quebec (without the accent) is the province's official name in Canadian English is one of the thirteen provinces and territories of Canada.

See Logistic regression and Quebec

Random variable

A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events.

See Logistic regression and Random variable

Rational choice theory

Rational choice theory refers to a set of guidelines that help understand economic and social behaviour.

See Logistic regression and Rational choice theory

Raymond Pearl

Raymond Pearl (June 3, 1879 – November 17, 1940) was an American biologist, regarded as one of the founders of biogerontology.

See Logistic regression and Raymond Pearl

Real number

In mathematics, a real number is a number that can be used to measure a continuous one-dimensional quantity such as a distance, duration or temperature.

See Logistic regression and Real number

Regression analysis

In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships between a dependent variable (often called the 'outcome' or 'response' variable, or a 'label' in machine learning parlance) and one or more independent variables (often called 'predictors', 'covariates', 'explanatory variables' or 'features').

See Logistic regression and Regression analysis

Regularization (mathematics)

In mathematics, statistics, finance, and computer science, particularly in machine learning and inverse problems, regularization is a process that changes the result answer to be "simpler".

See Logistic regression and Regularization (mathematics)

Ridge regression

Ridge regression is a method of estimating the coefficients of multiple-regression models in scenarios where the independent variables are highly correlated.

See Logistic regression and Ridge regression

Robust statistics

Robust statistics are statistics that maintain their properties even if the underlying distributional assumptions are incorrect.

See Logistic regression and Robust statistics

Ronald Fisher

Sir Ronald Aylmer Fisher (17 February 1890 – 29 July 1962) was a British polymath who was active as a mathematician, statistician, biologist, geneticist, and academic.

See Logistic regression and Ronald Fisher

Scalar (mathematics)

A scalar is an element of a field which is used to define a vector space.

See Logistic regression and Scalar (mathematics)

Scale parameter

In probability theory and statistics, a scale parameter is a special kind of numerical parameter of a parametric family of probability distributions.

See Logistic regression and Scale parameter

Science (journal)

Science, also widely referred to as Science Magazine, is the peer-reviewed academic journal of the American Association for the Advancement of Science (AAAS) and one of the world's top academic journals.

See Logistic regression and Science (journal)

Separation (statistics)

In statistics, separation is a phenomenon associated with models for dichotomous or categorical outcomes, including logistic and probit regression.

See Logistic regression and Separation (statistics)

Sigmoid function

A sigmoid function is any mathematical function whose graph has a characteristic S-shaped or sigmoid curve.

See Logistic regression and Sigmoid function

Softmax function

The softmax function, also known as softargmax or normalized exponential function, converts a vector of real numbers into a probability distribution of possible outcomes.

See Logistic regression and Softmax function

Sparse matrix

In numerical analysis and scientific computing, a sparse matrix or sparse array is a matrix in which most of the elements are zero.

See Logistic regression and Sparse matrix

Spline (mathematics)

In mathematics, a spline is a function defined piecewise by polynomials.

See Logistic regression and Spline (mathematics)

Stan (software)

Stan is a probabilistic programming language for statistical inference written in C++.

See Logistic regression and Stan (software)

Statistical classification

When classification is performed by a computer, statistical methods are normally used to develop the algorithm.

See Logistic regression and Statistical classification

Statistical data type

In statistics, groups of individual data points may be classified as belonging to any of various statistical data types, e.g. categorical ("red", "blue", "green"), real number, odd number (1,3,5) etc.

See Logistic regression and Statistical data type

Statistical model

A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of sample data (and similar data from a larger population).

See Logistic regression and Statistical model

Statistics

Statistics (from German: Statistik, "description of a state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data.

See Logistic regression and Statistics

Step function

In mathematics, a function on the real numbers is called a step function if it can be written as a finite linear combination of indicator functions of intervals.

See Logistic regression and Step function

Stratification (clinical trials)

Stratification of clinical trials is the partitioning of subjects and results by a factor other than the treatment given.

See Logistic regression and Stratification (clinical trials)

Supervised learning

Supervised learning (SL) is a paradigm in machine learning where input objects (for example, a vector of predictor variables) and a desired output value (also known as human-labeled supervisory signal) train a model.

See Logistic regression and Supervised learning

Trauma Quality Improvement Program

The Trauma Quality Improvement Program (TQIP) was initiated in 2008 by the American College of Surgeons Committee on Trauma.

See Logistic regression and Trauma Quality Improvement Program

Type I and type II errors

In statistical hypothesis testing, a type I error, or a false positive, is the rejection of the null hypothesis when it is actually true.

See Logistic regression and Type I and type II errors

Udny Yule

George Udny Yule, CBE, FRS (18 February 1871 – 26 June 1951), usually known as Udny Yule, was a British statistician, particularly known for the Yule distribution and proposing the preferential attachment model for random graphs.

See Logistic regression and Udny Yule

Unit of measurement

A unit of measurement, or unit of measure, is a definite magnitude of a quantity, defined and adopted by convention or by law, that is used as a standard for measurement of the same kind of quantity.

See Logistic regression and Unit of measurement

Utility

In economics, utility is a measure of the satisfaction that a certain person has from a certain state of the world.

See Logistic regression and Utility

Value (mathematics)

In mathematics, value may refer to several, strongly related notions.

See Logistic regression and Value (mathematics)

Variational Bayesian methods

Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning.

See Logistic regression and Variational Bayesian methods

Wald test

In statistics, the Wald test (named after Abraham Wald) assesses constraints on statistical parameters based on the weighted distance between the unrestricted estimate and its hypothesized value under the null hypothesis, where the weight is the precision of the estimate.

See Logistic regression and Wald test

Wilhelm Ostwald

Friedrich Wilhelm Ostwald (4 April 1932) was a Baltic German chemist and philosopher.

See Logistic regression and Wilhelm Ostwald

Y-intercept

In analytic geometry, using the common convention that the horizontal axis represents a variable x and the vertical axis represents a variable y, a y-intercept or vertical intercept is a point where the graph of a function or relation intersects the y-axis of the coordinate system.

See Logistic regression and Y-intercept

See also

Predictive analytics

Regression models

References

[1] https://en.wikipedia.org/wiki/Logistic_regression

Also known as Applications of logistic regression, Binary logit model, Conditional logit analysis, Logit model, Logit regression.

, Edwin Bidwell Wilson, Emergency management, Engineering, Entropy (information theory), Error function, Estimation theory, Euler numbers, Expectation propagation, Expected value, Exponential family, Exponential function, Generalized linear model, Goodness of fit, Gradient descent, Heavy-tailed distribution, Hosmer–Lemeshow test, Identifiability, Independence of irrelevant alternatives, Independent and identically distributed random variables, Indicator function, Information content, Iteratively reweighted least squares, Jane Worcester, Jarrow–Turnbull model, John Gaddum, Joseph Berkson, Journal of Clinical Epidemiology, Just another Gibbs sampler, Kullback–Leibler divergence, L. Gustave du Pasquier, Lagrange multiplier, Latent and observable variables, Latent variable model, Likelihood function, Likelihood-ratio test, Limited dependent variable, Limited-memory BFGS, Linear combination, Linear discriminant analysis, Linear function (calculus), Linear least squares, Linear model, Linear predictor function, Linear regression, Local case-control sampling, Location parameter, Logarithm, Logistic distribution, Logistic function, Logistic model tree, Logit, Loss function, Loss functions for classification, Lowell Reed, Luce's choice axiom, Machine learning, Mark Thoma, Marketing, Matching (statistics), Maximum a posteriori estimation, Maximum entropy probability distribution, Maximum likelihood estimation, Mixed logit, Mlpack, Mortgage, Multicollinearity, Multinomial logistic regression, Myocardial infarction, Natural language processing, Natural logarithm, Neural network (machine learning), Newton's method, Normal distribution, Normalizing constant, Null hypothesis, Null model, Observational study, Odds, Odds ratio, One in ten rule, OpenBUGS, Ordered logit, Ordinary least squares, Overfitting, Parti Québécois, Partition of sums of squares, Perceptron, Philosophical Transactions of the Royal Society A, Pierre François Verhulst, Poisson regression, Political science, Polynomial regression, Population growth, Posterior probability, Prior probability, Probabilistic programming, Probability, Probability distribution, Probability mass function, Probit, Probit model, Proceedings of the National Academy of Sciences of the United States of America, PyMC, Quasi-Newton method, Quebec, Random variable, Rational choice theory, Raymond Pearl, Real number, Regression analysis, Regularization (mathematics), Ridge regression, Robust statistics, Ronald Fisher, Scalar (mathematics), Scale parameter, Science (journal), Separation (statistics), Sigmoid function, Softmax function, Sparse matrix, Spline (mathematics), Stan (software), Statistical classification, Statistical data type, Statistical model, Statistics, Step function, Stratification (clinical trials), Supervised learning, Trauma Quality Improvement Program, Type I and type II errors, Udny Yule, Unit of measurement, Utility, Value (mathematics), Variational Bayesian methods, Wald test, Wilhelm Ostwald, Y-intercept.