nonparametric goodness-of-fit testing under gaussian modelsby yu i. ingster; irina a. suslina

3
Nonparametric Goodness-of-Fit Testing under Gaussian Models by Yu I. Ingster; Irina A. Suslina Review by: Christophe Pouet Journal of the American Statistical Association, Vol. 99, No. 466 (Apr., 2004), pp. 561-562 Published by: American Statistical Association Stable URL: http://www.jstor.org/stable/27590415 . Accessed: 14/06/2014 13:27 Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at . http://www.jstor.org/page/info/about/policies/terms.jsp . JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. . American Statistical Association is collaborating with JSTOR to digitize, preserve and extend access to Journal of the American Statistical Association. http://www.jstor.org This content downloaded from 195.34.79.253 on Sat, 14 Jun 2014 13:27:00 PM All use subject to JSTOR Terms and Conditions

Upload: review-by-christophe-pouet

Post on 20-Jan-2017

212 views

Category:

Documents


0 download

TRANSCRIPT

Nonparametric Goodness-of-Fit Testing under Gaussian Models by Yu I. Ingster; Irina A.SuslinaReview by: Christophe PouetJournal of the American Statistical Association, Vol. 99, No. 466 (Apr., 2004), pp. 561-562Published by: American Statistical AssociationStable URL: http://www.jstor.org/stable/27590415 .

Accessed: 14/06/2014 13:27

Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at .http://www.jstor.org/page/info/about/policies/terms.jsp

.JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range ofcontent in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new formsof scholarship. For more information about JSTOR, please contact [email protected].

.

American Statistical Association is collaborating with JSTOR to digitize, preserve and extend access to Journalof the American Statistical Association.

http://www.jstor.org

This content downloaded from 195.34.79.253 on Sat, 14 Jun 2014 13:27:00 PMAll use subject to JSTOR Terms and Conditions

Book Reviews

recent developments that move away from hypothesis tests toward model selec

tion and maximum likelihood, Bayesian, and approximate Bayesian inference

of parameters are absent. Although these topics are fairly new to population

genetics, empirical investigators who wish to answer the question posed in the

Preface?"given a collection of DNA sequences, what underlying forces are

responsible for the observed patterns of variability?"?would be interested to

know about them. Finally, although examples are used effectively to describe

individual methods, some synthesis that connects several techniques to the same

example dataset would help empirical population geneticists design analyses of

their own data.

Despite these limitations, if used in conjunction with texts that state results

in broader biological context, this book would be helpful to those with a math

ematical or statistical background who are encountering the subject for the first

time. The book is more theoretical than most books in the area, and because it is

more modern than related classics, such as those by Crow and Kimura (1970) and Ewens (1979), it will be an informative reference for researchers in the

field.

Noah A. Rosenberg

University of Southern California

REFERENCES

Crow, J. F., and Kimura, M. (1970), An Introduction to Population Genetics

Theory, Minneapolis: Burgess. Ewens, W. J. (1979), Mathematical Population Genetics. Berlin: Springer

Verlag. Hartl, D. L., and Clark, A. G. (1998), Principles of Population Genetics

(3rd ed.), Sunderland, MA: Sinauer.

Li, W.-H. (1997), Molecular Evolution, Sunderland, MA: Sinauer.

Nonlinear Estimation and Classification.

David D. Denison, Mark H. Hansen, Christopher C. Holmes, Bani Mallick, and Bin Yu (Eds.). New York: Springer-Verlag, 2003.

ISBN 0-387-95471-6. vii + 474 pp. $69.95 (P).

As an addition to Springer's series of Lecture Notes in Statistics, Nonlinear

Estimation and Classification is a compilation of 31 articles on recent develop ments on estimation and classification with nonlinear methods. The book grew out of a 2-week-long workshop of the same title held in spring of 2001 at the

Mathematical Sciences Research Institute in Berkeley, California. The organiz ers of the conference set a goal of promoting communication between several

disciplines that address issues in nonlinear estimation and classification, but of

ten with different emphases, terminology, and perspectives. Such disciplines in

clude machine learning, signal processing, image analysis, information theory,

optimization, applied mathematics, and of course statistics. The main charac

teristics and challenges that make these research topics exciting include high

dimensionality of the input variables, complex structure of the data, and the

need to process massive amounts of data. The traditional linear statistical tools

are generally incapable of handling such situations.

The articles are grouped as long papers and short papers. The 11 long pa

pers (with the average length of 20 pages) are from invited talks given by some

leading researchers, and the remaining 20 papers (with an average length of

11 pages) are from contributed talks that also address important questions in

the area. The papers are a mixture of review articles and presentations of new

research results.

Some of the long papers concentrate on specific areas of application or on

specific statistical issues; there are also review articles on general methodolo

gies. The papers that focus on a single area of application include natural image

modeling by wavelets (Choi and Baraniuk), scene labeling via a coarse-to

fine strategy (Geman), environmental monitoring using satellite images (Kiiveri et al.), freeway traffic-flow modeling (Bickel et al.), and internet traffic model

ing (Cao et al.). These works introduce complex modeling issues in real-world

applications. Marineantu and Dietterich compare two methods of estimating the conditional probability function in classification based on trees. Gray ad

dresses the problem of clustering Gaussian mixtures with quantization based

on information-theoretic tools.

Huang and Stone review extended linear modeling with tensor-product

splines, a general framework proposed and studied by Stone and collabora

tors. Theoretical, methodological, and computational aspects are touched on,

561

with a special emphasis on free-knot cases. Two exciting methods of clas

sification developed outside of statistics are reviewed, namely support-vector machines (SVMs) and boosting. Wahba et al. offer insight on SVMs from a tra

ditional statistical viewpoint, review adaptive tuning of SVMs, and also address

nonstandard SVM problems. Schapire, who pioneered the now-popular boost

ing method for improving a weak classifier, reviews its application in machine

learning. He covers the basics, error bounds, connections to game theory, lin

ear programming, and logistic regression. Some demonstrations are also given.

Mukherjee et al. examine the familiar regularization approach to regression,

especially from a stability standpoint, and derive useful error bounds.

Overall, I find the long papers quite readable, informative, and sufficiently

explained to provide a reasonably clear picture of the topics. One helpful as

pect is that efforts are made to explain things from the perspective of different

disciplines and connections (e.g., statistical views on boosting). The short papers cover a diverse range of issues, including risk bounds for

estimating a regression function or hazard function based on trees (or combin

ing trees) or neural networks of different kinds; considerations of prior assign ment for Bayesian hierarchical models or nonparametrics; data compression for remote sensing data; clustering in nonstandard contexts; wavelets or related

methods for categorical data, nonequispaced data, or high-frequency financial

data; confidence intervals for spline density estimation; longitudinal or growth curve modeling using mixed-effects MARS; clustering methods in bioinfor

matics; adaptive kernels for SVMs; ANOVA DDP models for a set of related

random distributions; logic regression using Boolean combinations of binary

covariates; and instability in nonlinear estimation and classification. Most of

these articles report some new results, but usually in a more or less sketchy way

(presumably due to space limitation), whereas a few others are mainly review

papers on a specific topic. Theoretical, methodological, and computational is

sues are considered.

Two notable features of the book are its wide coverage in content and its

self-contained chapters. This is, of course, not surprising given the background of the book. A consequence is that one can read the papers (called chapters in

the book) in any order according to personal interest.

The book's strength lies in the joint depiction by the active researchers

of the recent research works on nonlinear estimation and classification, some

from fields other than statistics. The reviews of the important (relatively) new

methodologies, some classes of applications, and computational aspects should

be quite accessible to a wide range of readers. The book is certainly worth read

ing for a researcher seriously interested in understanding the state-of-the-art

methods and potential applications in high-dimensional estimation and classi

fication. I also think that the book can serve well for teaching a topics course

on modern estimation and classification in which advanced graduate students

select a topic for study and presentation. Of course, this book is not meant to be exhaustive or a perfect representation

of the recent and the current works under the theme of nonlinear estimation and

classification, nor is it sufficiently detailed for an in-depth understanding of the

technical issues. Indeed, in most cases the resolution levels of the materials are

relatively low, and the reader has the opportunity to use the references provided in the articles for further reading.

Nonlinear estimation and classification is an important and rapidly growing research area that has benefited from several disciplines. From this book, it is

clear that many challenges and opportunities remain for future research.

Yuhong Yang

Iowa State University

Nonparametric Goodness-of-Fit Testing Under Gaussian Models.

Yu I. INGSTER and Irina A. SUSLINA. New York: Springer-Verlag. 2003.

ISBN 0-387-95531-3. xiv + 452 pp. $98.00 (P).

This book concerns nonparametric goodness-of-fit testing problems. The

chosen approach is the minimax asymptotic approach (small noise level or

large sample size). The authors explain the choice of the minimax tests

by first studying the classical goodness-of-fit tests (Kolmogorov-Smirnov, Cram?r-von Mises). They also highlight the similarities and differences be

tween estimation and testing. The most modern aspects are studied; including

adaptive problems, multichannel detection, and composite null hypotheses. The

This content downloaded from 195.34.79.253 on Sat, 14 Jun 2014 13:27:00 PMAll use subject to JSTOR Terms and Conditions

562

book is self-contained, and the bibliography is very rich and in fact provides a comprehensive listing of references about minimax testing (something that

heretofore had been missing from the field). The authors explain the minimax approach and its connection with the

Bayesian approach. There are two kind of problems: the sharp asymptotic prob lem and the simpler problem about distinguishability and undistinguishability conditions.

To get the best out of this book, the reader should be familiar with basic func

tional analysis, wavelet theory, and optimization for extreme problems. There

are only a few typographical errors, and they can easily be corrected by the

reader. This book is intended for graduate students and researchers. Parts of

this book can serve as starting points for a course at a graduate level. It is highly recommended to anyone who wants an introduction to hypothesis testing from

the minimax approach?yet it is only a starting point, as Gaussian models are

studied exclusively. For other models, the reader will have to get articles or wait

for another book on density or spectral density models, maybe that by Ingster and Suslina.

The authors begin with the distinguishability problem. Their purpose is to in

troduce the finite-dimensional and infinite-dimensional problems. They clearly

explain the need for smoothness assumptions, especially the importance of the

norm to define the removed neighborhood. They make the connection between

sequence spaces (power norms and Besov norms are considered) and functional

spaces (Besov and Sobolev spaces) through the wavelet theory. Differences between estimation and hypothesis testing are indicated. The

testing problem is important because the plug-in estimation test does not pro vide a good rate of convergence in nonclassical results. Conditions for distin

guishability and undistinguishability are given for ellipsoids. The concept of

semiorthogonal priors is also introduced.

The tools introduced for the study of the lower bound are very important and

are applied systematically in the case of power norm, Besov norm, and positive alternatives. An interesting point that the authors make is about the Besov norm,

the parameters can lead to trivial results.

Next, the sharp asymptotics problem is considered, and the links between

the upper bound and the lower bound are clearly explained. This leads to a very

general problem of optimization. This approach is very demanding, but gives a deep understanding of the results. A general approach is given to justify the

choice of two- and three-point priors. Sharp asymptotics lead to an optimization

problem on the space of probability measures.

In the infinite-dimensional model with the level noise fading, three types of

results are obtained: trivial, asymptotics of degenerate types, and sharp asymp totics of Gaussian types.

The last part of the book is devoted to more specialized and modern topics. The problem of adaptation, introduced in testing by Spokoiny in 1996, is con

sidered here in the case of the power norm and Besov norm. The second type of

result involves high-dimensional signal detection. Here new kinds of asymptot ics with infinite-divisible distributions are obtained. This problem is related to

multichannel signal detection. The third modern topic is the problem of testing a composite null hypothesis, especially a hypothesis that is close to a simple

hypothesis. The rates of testing are the same as in the simple null hypothesis

counterparts. Yet, the size of the composite null hypothesis depends heavily on

the distance that measures this size (Lp norm).

Christophe Pouet

Universit? de Provence

REFERENCES

Lepski, O. V, Nemirovsky, A., and Spokoiny, V G. (1999), "On Estimation

of the Lr Norm of a Regression Function," Probability Theory and Related

Fields, 113, 221-253.

Spokoiny, V. G. (1996), "Adaptive Hypothesis Testing Using Wavelets," The

Annals of Statistics, 24, 2477-2498.

Applied Probability.

Kenneth Lange. New York: Springer-Verlag, 2003. ISBN 0-387

00425-4. xii + 300 pp. $79.95 (H).

Professor Lange has presented us with a graduate-level text in probability. In his Preface, he worries that the pursuit of mathematical rigor discourages students of science, particularly of biology, from learning the powerful tools

Book Reviews

that modern probability theory puts at their disposal. This book is an attempt to remedy that. The first chapter is a brief, quick introduction to mathemati

cal probability, with almost no proofs. I believe this chapter should constitute

a warning to students of the sorts of things they should already have seen in

their education, that they may be called on to remember and use in what fol

lows. An introductory graduate, or perhaps upper-level undergraduate, course

in mathematical probability should usually suffice.

The rest of the book is a series of chapters that introduce probabilistic tools,

with many illustrative applications, to mathematical biology (particularly ge

netics) and other branches of mathematics. The author's professional interests

certainly guided his choices, but the selection of methods and areas of applica tion is nevertheless so broad that readers with other specialties will likely find

their curiosity piqued. This manner of development meant that I often would have chosen alterna

tive approaches to the applications; but this seems beside the point. The author

wanted to show the breadth and power of the tools. Several times I was pleas

antly surprised to learn that an application could be tackled in an unfamiliar

way. Because the book is comparatively short, it can by no means be a survey

of all the important tools of modern applied probability. For example, little is

given of asymptotic approximation theory, and comparatively little of stochas

tic processes. Nor are the tools that do appear studied to great depth. But the

ones that are covered seem well chosen and are well explained. More than once

I found informative discussions of state-of-the-art techniques that I had recently encountered in the research literature and wished to learn more about. The book

is rich with pointers to an excellent bibliography for those wishing to pursue

subjects further.

Chapter 2 covers devices for finding expectations, including familiar ones

like indicators and generating functions and less-familiar tricks using condi

tioning. Chapter 3 is an entertainingly idiosyncratic chapter on convexity, with

some standard connections to inequalities, and less familiar ones to optimiza tion problems.

Chapters 4 and 5 consider combinatorics, at an intermediate level beyond that known to every student of probability. The techniques exploited include

inclusion-exclusion and the pigeonhole principle. Applications include an in

teresting introduction to probabilistic techniques for evaluating algorithms.

Chapter 6 introduces Poisson processes, with applications to such fields as

tomography. There are several illustrations of what the author calls Poissoniza

tion; let an integer parameter be a Poisson random variable, then condition on

its value. Later, Chapter 12 introduces some rather modern techniques in Pois

son approximation, such as the Chen-Stein bounds.

Chapters 7 and 8 introduce the classic topics of discrete and continuous

Markov chains. Chapter 9 provides a nice introduction to branching processes. All of these chapters present a number of biological applications, such as to

genetics and epidemiology.

Chapter 10, on martingales, will no doubt please statisticians for its applica tion to sequential testing, but will equally certainly frustrate specialists for its

brevity. Chapter 11, on diffusion processes, is surprisingly rich in theory and

applications even at this fairly elementary level. Chapter 13 is an unexpected

chapter, dealing with probabilistic techniques in number theory. It climaxes

with a proof of the prime number theorem, of all things. Professor Lange's book is certainly a pleasure to read, and it is written in a

clear style using standard probabilistic notation. Each chapter ends with exer

cises ranging from moderate to challenging in difficulty. These exercises serve

to fill in mathematical details, elaborate on the applications discussed, and sug

gest further applications. I expect to keep this book handy, because my interests require that I be able

to resurrect knowledge of some of the many techniques found within. The au

thor suggests that the book may be useful in teaching graduate students in the

sciences the probability that they need for their research. In my experience, very few graduate students in the sciences are ready for the mathematical level re

quired here. Rather, the book perhaps could be used as the text for the second

part of a serious graduate course in probability for an applied probability, statis

tics, or mathematics program. The first part of the course would be taught from

a good mathematical probability textbook, which this book clearly presupposes.

George R. TERRELL

Virginia Polytechnic Institute and State University

This content downloaded from 195.34.79.253 on Sat, 14 Jun 2014 13:27:00 PMAll use subject to JSTOR Terms and Conditions