TY - CHAP
T1 - 30 Sequential methods of analysis for genome scans
AU - Province, Michael A.
N1 - Funding Information:
This work was partly supported by National Institutes of Health grants GM28719 from the National Institute of General Medical Sciences, and HL54473 from the National Heart, Lung, and Blood Institute. The S.A.G.E. results were obtained by using the program package S.A.G.E., which is supported by a U.S. Public Health Service resource grant (1 P41 RRO3655) from the National Center for Research Resources.
PY - 2001
Y1 - 2001
N2 - As the preceding chapters illustrate, now that whole-genome scan analyses are becoming more common, there is considerable disagreement about the best way to balance between false positives and false negatives (traditionally called type I and type II errors in the statistical parlance). Type I and type II errors can be simultaneously controlled, if we are willing to let the sample size of analysis vary. This is the secret that Wald 1947 discovered in the 1940s that led to the theory of sequential sampling and was the inspiration for Newton Morton in developing the lod score method. We can exploit this idea further and capitalize on an old, but nearly forgotten theory: sequential multiple decision procedures (SMDP) Bechhoffer, et al. 1968, which generalizes the standard "two-hypotheses" tests to consider multiple alternative hypotheses. Using this theory, we can develop a single, genome-wide test that simultaneously partitions all markers into "signal" and "noise" groups, with tight control over both type I and type II errors (Province, 2000). Conceiving this approach as an analysis tool for fixed sample design (instead of a true sequential sampling scheme), we can let the data decide at which point we should move from the hypothesis generation phase of a genome scan (where multiple comparisons make the interpretation of p values and significance levels difficult and controversial), to a true hypothesis-testing phase (where the problem of multiple comparison of multiple comparison has been all but eliminated so that p values may be accepted at face value.
AB - As the preceding chapters illustrate, now that whole-genome scan analyses are becoming more common, there is considerable disagreement about the best way to balance between false positives and false negatives (traditionally called type I and type II errors in the statistical parlance). Type I and type II errors can be simultaneously controlled, if we are willing to let the sample size of analysis vary. This is the secret that Wald 1947 discovered in the 1940s that led to the theory of sequential sampling and was the inspiration for Newton Morton in developing the lod score method. We can exploit this idea further and capitalize on an old, but nearly forgotten theory: sequential multiple decision procedures (SMDP) Bechhoffer, et al. 1968, which generalizes the standard "two-hypotheses" tests to consider multiple alternative hypotheses. Using this theory, we can develop a single, genome-wide test that simultaneously partitions all markers into "signal" and "noise" groups, with tight control over both type I and type II errors (Province, 2000). Conceiving this approach as an analysis tool for fixed sample design (instead of a true sequential sampling scheme), we can let the data decide at which point we should move from the hypothesis generation phase of a genome scan (where multiple comparisons make the interpretation of p values and significance levels difficult and controversial), to a true hypothesis-testing phase (where the problem of multiple comparison of multiple comparison has been all but eliminated so that p values may be accepted at face value.
UR - http://www.scopus.com/inward/record.url?scp=0035219882&partnerID=8YFLogxK
U2 - 10.1016/s0065-2660(01)42039-6
DO - 10.1016/s0065-2660(01)42039-6
M3 - Chapter
C2 - 11037338
AN - SCOPUS:0035219882
SN - 0120176424
SN - 9780120176427
T3 - Advances in Genetics
SP - 499
EP - 514
BT - Advances in Genetics
PB - Academic Press Inc.
ER -