Sequential methods of analysis for genome scans

Adv Genet. 2001:42:499-514. doi: 10.1016/s0065-2660(01)42039-6.

Abstract

As the preceding chapters illustrate, now that whole-genome scan analyses are becoming more common, there is considerable disagreement about the best way to balance between false positives and false negatives (traditionally called type I and type II errors in the statistical parlance). Type I and type II errors can be simultaneously controlled, if we are willing to let the sample size of analysis vary. This is the secret that Wald (1947) discovered in the 1940s that led to the theory of sequential sampling and was the inspiration for Newton Morton in developing the lod score method. We can exploit this idea further and capitalize on an old, but nearly forgotten theory: sequential multiple decision procedures (SMDP) (Bechhoffer, et al., 1968), which generalizes the standard "two-hypotheses" tests to consider multiple alternative hypotheses. Using this theory, we can develop a single, genome-wide test that simultaneously partitions all markers into "signal" and "noise" groups, with tight control over both type I and type II errors (Province, 2000). Conceiving this approach as an analysis tool for fixed sample designs (instead of a true sequential sampling scheme), we can let the data decide at which point we should move from the hypothesis generation phase of a genome scan (where multiple comparisons make the interpretation of p values and significance levels difficult and controversial), to a true hypothesis-testing phase (where the problem of multiple comparisons has been all but eliminated so that p values may be accepted at face value).

Publication types

  • Comparative Study
  • Research Support, U.S. Gov't, P.H.S.
  • Review

MeSH terms

  • Chromosome Mapping / methods*
  • Computer Simulation
  • Genetic Linkage*
  • Genome, Human*
  • Humans
  • Monte Carlo Method
  • Quantitative Trait, Heritable