A statistical model is a collection of probability distributions on some sample space. We assume that the collection, 𝒫, is indexed by some set Θ. The set Θ is called the parameter set or, more commonly, the parameter space. For each θ ∈ Θ, let Fθ denote the corresponding member of the collection; so Fθ is a cumulative distribution function. Then a statistical model can be written as
The model is a parametric model if Θ ⊆ ℝk for some positive integer k.
When the model consists of absolutely continuous distributions, it is often specified in terms of corresponding probability density functions:
where pλ is the probability mass function. This family is an exponential family.
This parametrized family is both an exponential family and a location-scale family.
This example illustrates the definition for a model with some discrete parameters.
A parametric model is called identifiable if the mapping θ ↦ Pθ is invertible, i.e. there are no two different parameter values θ1 and θ2 such that Pθ1 = Pθ2.
Parametric models are contrasted with the semi-parametric, semi-nonparametric, and non-parametric models, all of which consist of an infinite set of "parameters" for description. The distinction between these four classes is as follows:
Some statisticians believe that the concepts "parametric", "non-parametric", and "semi-parametric" are ambiguous.1 It can also be noted that the set of all probability measures has cardinality of continuum, and therefore it is possible to parametrize any model at all by a single number in (0,1) interval.2 This difficulty can be avoided by considering only "smooth" parametric models.
Le Cam & Yang 2000, §7.4 - Le Cam, Lucien; Yang, Grace Lo (2000), Asymptotics in Statistics: Some basic concepts (2nd ed.), Springer ↩
Bickel et al. 1998, p. 2 - Bickel, Peter J.; Klaassen, Chris A. J.; Ritov, Ya’acov; Wellner, Jon A. (1998), Efficient and Adaptive Estimation for Semiparametric Models, Springer ↩