In statistics, the method of estimating equations is a way of specifying how the parameters of a statistical model should be estimated. This can be thought of as a generalisation of many classical methods—the method of moments, least squares, and maximum likelihood—as well as some recent methods like M-estimators.
The basis of the method is to have, or to find, a set of simultaneous equations involving both the sample data and the unknown model parameters which are to be solved in order to define the estimates of the parameters. Various components of the equations are defined in terms of the set of observed data on which the estimates are to be based.
Important examples of estimating equations are the likelihood equations.
Examples [ edit ]
while the estimating equation based on the median is
Each of these equations is derived by equating a sample value (sample statistic) to a theoretical (population) value. In each case the sample statistic is a consistent estimator of the population value, and this provides an intuitive justification for this type of approach to estimation.
See also [ edit ]
- Generalized estimating equations
- Method of moments (statistics)
- Generalized method of moments
- Maximum likelihood
References [ edit ]
- Godambe, V. P., ed. (1991). Estimating Functions. New York: Oxford University Press. ISBN 0-19-852228-2.
- Heyde, Christopher C. (1997). Quasi-Likelihood and Its Application: A General Approach to Optimal Parameter Estimation. New York: Springer-Verlag. ISBN 0-387-98225-6.
- McLeish, D. L.; Small, Christopher G. (1988). The Theory and Applications of Statistical Inference Functions. New York: Springer-Verlag. ISBN 0-387-96720-6.
- Small, Christopher G.; Wang, Jinfang (2003). Numerical Methods for Nonlinear Estimating Equations. New York: Oxford University Press. ISBN 0-19-850688-0.