# Generative model

## Model for generating observable data in probability and statistics / From Wikipedia, the free encyclopedia

#### Dear Wikiwand AI, let's keep it short by simply answering these key questions:

Can you list the top facts and stats about Generative model?

Summarize this article for a 10 year old

In statistical classification, two main approaches are called the **generative** approach and the **discriminative** approach. These compute classifiers by different approaches, differing in the degree of statistical modelling. Terminology is inconsistent,^{[lower-alpha 1]} but three major types can be distinguished, following Jebara (2004):

- A
**generative model**is a statistical model of the joint probability distribution $P(X,Y)$ on given observable variable*X*and target variable*Y*;^{[1]} - A
**discriminative model**is a model of the conditional probability $P(Y\mid X=x)$ of the target*Y*, given an observation*x*; and - Classifiers computed without using a probability model are also referred to loosely as "discriminative".

The distinction between these last two classes is not consistently made;^{[2]} Jebara (2004) refers to these three classes as *generative learning*, *conditional learning*, and *discriminative learning*, but Ng & Jordan (2002) only distinguish two classes, calling them **generative classifiers** (joint distribution) and **discriminative classifiers** (conditional distribution or no distribution), not distinguishing between the latter two classes.^{[3]} Analogously, a classifier based on a generative model is a **generative classifier**, while a classifier based on a discriminative model is a **discriminative classifier**, though this term also refers to classifiers that are not based on a model.

Standard examples of each, all of which are linear classifiers, are:

- generative classifiers:
- discriminative model:

In application to classification, one wishes to go from an observation *x* to a label *y* (or probability distribution on labels). One can compute this directly, without using a probability distribution (*distribution-free classifier*); one can estimate the probability of a label given an observation, $P(Y|X=x)$ (*discriminative model*), and base classification on that; or one can estimate the joint distribution $P(X,Y)$ (*generative model*), from that compute the conditional probability $P(Y|X=x)$, and then base classification on that. These are increasingly indirect, but increasingly probabilistic, allowing more domain knowledge and probability theory to be applied. In practice different approaches are used, depending on the particular problem, and hybrids can combine strengths of multiple approaches.