@ARTICLE\{IMM2002-01445, author = "S. C. Strother and J. Anderson and L. K. Hansen", title = "The Quantitative Evaluation of Functional Neuroimaging Experiments: The {NPAIRS}", year = "2002", keywords = "{NPAIRS,} neuroimaging, modeling", pages = "747-771", journal = "NeuroImage", volume = "15", editor = "", number = "4", publisher = "", url = "http://www2.compute.dtu.dk/pubdb/pubs/1445-full.html", abstract = "We introduce a data-analysis framework and performance metrics for evaluating and optimizing the interaction between activation tasks, experimental designs, and the methodological choices and tools for data acquisition, preprocessing, data analysis, and extraction of statistical parametric maps (SPMs). Our {NPAIRS} (nonparametric prediction, activation, influence, and reproducibility resampling) framework provides an alternative to simulations and {ROC} curves by using real {PET} and fMRI data sets to examine the relationship between prediction accuracy and the signal-to-noise ratios (SNRs) associated with reproducible SPMs. Using cross-validation resampling we plot training-test set predictions of the experimental design variables (e.g., brain-state labels) versus reproducibility {SNR} metrics for the associated SPMs. We demonstrate the utility of this framework across the wide range of performance metrics obtained from \$\backslash\$[/sup 15/ O]water {PET} studies of 12 age- and sex-matched data sets performing different motor tasks (8 subjects/set). For the 12 data sets we apply {NPAIRS} with both univariate and multivariate data-analysis approaches to: (1) demonstrate that this framework may be used to obtain reproducible SPMs from any data-analysis approach on a common Z -score scale (rSPM\{ Z \}); (2) demonstrate that the histogram of a rSPM\{ Z \} image may be modeled as the sum of a data-analysis-dependent noise distribution and a task-dependent, Gaussian signal distribution that scales monotonically with our reproducibility performance metric; (3) explore the relation between prediction and reproducibility performance metrics with an emphasis on bias-variance tradeoffs for flexible, multivariate models; and (4) measure the broad range of reproducibility SNRs and the significant influence of individual subjects. A companion paper describes learning curves for four of these 12 data sets, which describe an alternative mutual-information prediction metric and {NPAIRS} reproducibility as a function of training-set sizes from 2 to 18 subjects. We propose the {NPAIRS} framework as a validation tool for testing and optimizing methodological choices and tools in functional neuroimaging" }