OPTIMAL PLUG-IN ESTIMATORS FOR NONPARAMETRIC FUNCTIONAL ESTIMATION
成果类型:
Article
署名作者:
GOLDSTEIN, L; MESSER, K
署名单位:
California State University System; California State University Fullerton
刊物名称:
ANNALS OF STATISTICS
ISSN/ISSBN:
0090-5364
DOI:
10.1214/aos/1176348770
发表日期:
1992
页码:
1306-1328
关键词:
squared density derivatives
CONVERGENCE
摘要:
Consider the problem of estimating the value of a functional LAMBDA(f) for f an unknown density or regression function. The straightforward plug-in estimator LAMBDA(f) with f a particular estimate of f achieves the optimal rate of convergence in the sense of Stone over bounded subsets of a Sobolev space for a broad class of linear and nonlinear functionals. For many functionals the rate calculation depends on a Frechet-like derivative of the functional, which may be obtained using elementary calculus. For some classes of functionals, f is undersmoothed relative to what would be used to estimate f optimally. Examples for which a plug-in estimator is optimal include L(q) norms of regression or density functions and their derivatives and the expected integrated squared bias. When interested in computing estimates over classes of functions which satisfy certain restrictions, such as strict positivity or boundary conditions, the plug-in estimator may or may not be optimal, depending on the functional and the function class. The functional calculus establishes conditions under which the plug-in estimator remains optimal, and sometimes suggests an appropriate modification when it does not.