Knowledge

Talk:Multiple comparisons problem

Source šŸ“

1139:
certain defined conditions"). Also, the first category (methods that work "under any conditions") is so narrowly defined that it excludes basically any standard method, as there are pretty much always statistical assumptions that are required. And the second category (methods that work "except under certain defined conditions") is so vague as to be essentially meaningless. Also, some of the info here is false. For example, Tukey's range test is NOT an omnibus test; and a method can require a significant omnibus test and still provide strong control (e.g., Hayter's procedure). Also, the list of categories ignores procedures that are designed to control the false discovery rate, rather than the familywise error rate. It would be better to link to the familywise error rate article than attempt to cram a bunch of detail in here.
667:, by Westfall, Tobias, and Wolfinger (all respected names in the field), the authors say "the term 'multiple testing' is more common than 'multiple comparisons' when analyzing modern high-dimensional data." They make a fairly loose distinction between "multiple comparisons" and "multiple testing," saying that "multiple comparisons" is "often" used for comparisons of several treatment means, whereas "multiple testing" is used in "a broader class of applications." But certainly if one substituted nonparametric tests (in which case the comparisons wouldn't be of means), there would still be "multiple comparisons" in the literal sense. In fact, even multiple correlation tests can be considered multiple "comparisons," as each test is effectively a comparison of the correlation to zero. 641:
of P-values is the most common method for doing this, but not the only one. Multiple comparisons are the special cases where the tests are comparisons between groups, typically several pairwise comparisons on a set of groups (e.g. all-against-all or one-against-all); although Bonferroni correction is still valid, it is typically not very powerful as it does not make use of the dependencies between different pairwise tests performed on the same groups. I'll see if I can make an update to the multiple testing page at some point which will make this difference more clear. --
656:, Shaffer, 1995). In various instances over the decades, people have sometimes used the term "multiple comparisons" more narrowly (referring to ANOVA-type contexts), but I disagree with Septagon's statement that "multiple comparisons" can only be used in the context of group comparisons. For example, in a within-subjects (repeated measures) design, there aren't separate groups, but there could still be "multiple comparisons" if there are more than one factor or more than two factor levels or more than one outcome variable. 85: 64: 1133:
except under certain defined conditions. Methods which rely on an omnibus test before proceeding to multiple comparisons. Typically these methods require a significant ANOVA, MANOVA, or Tukey's range test. These methods generally provide only "weak" control of Type I error, except for certain numbers of hypotheses. Empirical methods, which control the proportion of Type I errors adaptively, utilizing correlation and distribution characteristics of the observed data.
806:"If the inferences are hypothesis tests rather than confidence intervals, the same issue arises. With just one test performed at the 5% level, there is only a 5% chance of incorrectly rejecting the null hypothesis if the null hypothesis is true. However, for 100 tests where all null hypotheses are false, the expected number of incorrect rejections is 5. If the tests are independent, the probability of at least one incorrect rejection is 99.4%. These errors are called 1189:
distributional assumptions" is vague and likely unclear to many readers. I suggest either being more explicit (e.g., "The Bonferroni procedure does not require any assumptions about the dependence (correlation) of the tests, and does not impose any added assumptions about the distributions of the test statistics") or simply linking to the Bonferroni correction article without going into the issue of assumptions here.
159: 382:", then indeed we are not talking about multiple testing in general (which covers more generally the problems of "using statistical tests repeatedly" as indicated in the intro). Do you agreeĀ ? In this case, most of the discussion could go in a more general article on multiple testing, that I will be happy to start. Unfortunately, all the definitions I have been able to find so far for 33: 1198:
pros and cons, and some procedures are more generalized or specialized than others. Mentioning Bonferroni makes sense (though we don't need to do so multiple times in the article), since that's the simplest method. Why don't we just mention Bonferroni and note that there are other procedures, rather than going into detail about two particular alternatives?
551:
or wilcoxon, it is more likely having false positives (the coin example in the article). In this way, we would be accepting a false hipotesis for example saying that A has the same mean that D. For this reason, we have tests designed to avoid this: ANOVA (parametric), Friedman (nonparametric), others??...
901:
The table is confusing, mostly because there is no clear relation between rows and columns. I believe "Declared significant" means "Researchers believe alternative hypothesis to be true", and "Declared non-significant" means "Researchers believe null hypothesis to be true". This relabeling might make
498:
about multiple testing (starting with Benjamini-Hochberg, as mentioned above). This is probably why I easily believed that multiple comparisons was the special case, but it may be only a bias. In any case, if the consensus is that multiple testing==multiple comparison (hopefully other people will say
1197:
That method (i.e., the Å idĆ”k procedure) is indeed marginally less conservative than Bonferroni, but it's actually marginally liberal when the tests are negatively dependent. Also, focusing on the Å idĆ”k and Holm procedures is abritrary. Numerous proceudres have been developed. Each procedures has its
1089:
In some situations, the Bonferroni correction is substantially conservative, i.e., the actual family-wise error rate is much less than the prescribed level Ī±. This occurs when the test statistics are highly dependent (in the extreme case where the tests are perfectly dependent, the family-wise error
640:
should remain different pages. The difference is not merely UK/US or terminology, but one of essential difference. Multiple testing, or multiple hypotheses testing, is the general problem of testing several null hypotheses while controlling the overall chance of false positive; Bonferroni correction
1132:
Methods where total alpha can be proved to never exceed 0.05 (or some other chosen value) under any conditions. These methods provide "strong" control against Type I error, in all conditions including a partially correct null hypothesis. Methods where total alpha can be proved not to exceed 0.05
554:
After performing ANOVA or Friedman, we only know that for example H0: A = B = C = D is not true. Then we would probably want to know which one is different from the others. For this purpose, we can apply one of the techniques that allow us to _compare_ every pair: Tukey test, Nemenyi, Bonferroni...
550:
With these definitions, I think that making _multiple test_ is repeating a test some times. An example, if we want to test if A is better than B (or equal, or whatever). After that we got C and we want to test A vs C, and B vs C. Then comes D and I want A-D, B-D, C-D... If we do that, with a t-test
493:
conception, I was ok with the distinction and spinned-off the multiple testing article, which no one objected about. This is why I am a bit puzzled about the going back. I wonder if there may be a systematic difference in vocabulary between statisticians working in different fields; the statistical
1188:
Not necessarily. There are procedures that can be more conservative than Bonferroni in some cases (e.g., the Benjaminiā€“Yekutieli procedure and the ScheffĆ© procedure). Also, note that we don't need to mention the Bonferroni procedure both here and in the next section. Also, "free of dependence and
1143:
The advent of computerized resampling methods, such as bootstrapping and Monte Carlo simulations, has given rise to many techniques in the latter category. In some cases where exhaustive permutation resampling is performed, these tests provide exact, strong control of Type I error rates; in other
467:
address the multiple comparisons problem. They alas do not have an entry on "multiple testing". It seems like the article text (parts not by me) and all of the statistical tests linked below that I am familiar with address "multiple comparisons" as the problem is conceived by me and Mathworld. Is
1148:
That may be true (though no source is provided). But again, there is an entire wiki article on the familywise error rate, which covers methods of control. Why not just link to that article, instead of going into detail in this article which is supposed to be about the general problem of multiple
576:
Multiple Hypotheses: the analysis of variance. This chapter introduces the problem of repeating the same test to over different samples to confirm various hypothesis over them (coin example). Then explains ANOVA and their non-paramentric extensions like Kruskal Walls and points to chapter 14 for
445:
really refers to the application of multiple statistical tests, whatever they are. The google searches do not tell us if the two terms have the exact same meaning (some of the links for multiple comparisons point towards the ANOVA question only; some talk about the more general problem). For the
1138:
That entire list is unsourced, and there is no particular reason to divide multiple testing procedures into those particular categories. For one thing, these categories are not mutually exclusive (e.g., methods that provide weak control are a subset of methods that provide control "except under
622:
I'm late to this discussion, but I've never heard of multiple testing, until just now. I use multiple comparisons as a term all the time (especially Bonferroni and friends). Could it be a UK/US thing, or a case where SPSS has dictated the vocabulary to the world? Otherwise, I think the time for
1076:
Not true. For instance, you could use sequential testing or a closed testing procedure. Also, the familywise error rate is only one "total error rate" that has been defined. It would be better to mention the familywise error rate and false discovery rate, and link to the wiki articles on those
897:
The section comes at an important point in the article and risks throwing the reader off-track. The section throws a lot of variables at the reader, forcing him or her to ponder a complicated table just to come up with some pretty intuitive and obvious ideas, such as that the number of false
651:
I'm late to the party here, but "multiple comparisons" and "multiple testing" are typically considered as fairly interchangeable terms these days. One of the most respected experts on the topic is J.P. Shaffer, who said: "The term 'multiple comparisons' has come to be used synonymously with
948:
The article is not structured well, has many holes, and even contains some wrong (or at least misleading) statements. It might be better to scrap it altogether and refer to a well-crafted review paper instead. Sorry about this negative rating but I feel the need to say it as it is.
558:
The previous could clearly split article in two, but probably I have left other ideas, like those about techniques to repeat a test in order to increase power that I do not know of. I think we should clearify which contents do we want here before deciding about one or two articles.
699:
I agree. I don't know what the sentence "The experimentwise Ī± level increases exponentially as the number of comparisons increases." means. What is an Ī± level, or where do I go to look it up? Not really a field I know that much about, so I look forward to a more clear article. --
992:
My issues with this article are: 1) the controlling procedures section is a repetitive since its merge with multiple testing correction so that needs to be condensed and cleaned up 2) the large-scale multiple testing section is largely missing citations for its claims
1119:
Because simple techniques such as the Bonferroni method can be conservative, there has been a great deal of attention paid to developing better techniques, such that the overall rate of false positives can be maintained without excessively inflating the rate of false
357:
use this term (I just added a sentence on this on the microarray page). Given that noone else has answered my suggestion (thanks for jumping inĀ !), I will not move the page, but I will indicate clearly that multiple testing is (also) what we are talking about here.
288:
Any outcome of tossing a fair coin 10 times has probibility of (1/2) ^ 10. There is only one way to get 10 heads and then 10 ways of getting 9 heads (once for each of the 10 throws being a tail e.g. first is a tail or third is a tail) hence (1 + 10) * (1/2).
659:"Multiple comparisons" is an older term than "multiple testing," so it is not surprising that it shows up more in searches. But "multiple testing" is now very much a standard, well accepted term in the field. For instance, there is a well known text called 403:. I have used it myself, and had reviewers of my own research papers claim I need to do "multiple comparisons corrections." Also, google returns 177,000 hits for the search bonferroni+"multiple comparisons", and 48,000 for bonferroni+"multiple testing". 1016:), would still need to have a source brought for it and deleted otherwise. Well here are the sources, you could just google "MCP conference XXX" (when XXX is the year of the confernce) and get these as first results, but I'll list those results anyway: 854: 820:
that is, we reject it when we think is false (based on the alpha level). In this case the problem arise because we reject it even if it is true. Sorry if I misunderstood- non-statistician here.Diego Diez 13:22, 23 September 2010 (UTC)
468:
your conception of multiple comparisons (i.e. the ANOVA f-test) a specific example of multiple testing/my conception of multiple comparisons? I wonder if we aren't just running into a linguistic rather than content-based hurdle here.
1158:
If m independent comparisons are performed, the family-wise error rate (FWER), is given by Ī±ĀÆ = 1 āˆ’ ( 1 āˆ’ Ī±)^m . Hence, unless the tests are perfectly positively dependent (i.e., identical), Ī±ĀÆ increases as the number of comparisons
861: 163: 1067:
Not clear what "recalculating probabilities" means. And "repeated multiple times" doesn't make sense in this context. Multiple testing doesn't mean the same test is repeated; it means multiple different tests are conducted.
1094:
Too much detail about one specific procedure (the Bonferroni procedure) in one implausible situation (perfect dependence). Why not just briefly mention the Bonferroni procedure and link to the wiki article on that topic?
499:
something), then the first priority would be to merge the other article, instead of rewriting it (although it may be too late). As a side note, at least some of the linked articles are indeed specially related to ANOVA.
893:
I suggest a re-thinking of the purpose and goal of this section. The main problem is that nothing in this section appears elsewhere in the article, so it should be deleted if no more work is done to make it useful.
419:
multiple comparison problem; indeed, most of the procedures linked from this page are specifically about "comparing sets of means", as was the lead section. Based on this, the article has been split between
1163:
The second statement, though true, doesn't follow from the first, so why "Hence?" Also, FWER is only one error rate that's been defined; what about the per-family error rate and the false discovery rate?
670:
To avoid confusion, maybe the best approach for this article is treat "multiple comparisons" as indeed equivalent to "multiple testing," but mention that the term has sometimes been defined more narrowly.
970:: Joseph P. Romano & Azeem M. Shaikh & Michael Wolf, 2009. "Hypothesis testing in econometrics," IEW - Working Papers 444, Institute for Empirical Research in Economics - University of Zurich. 437:
the general problem, then the two pages should be merged ā€” and this article should be cleaned up. But I must say that I like the split approach, and it seems logical: with the ANOVA, you are really
135: 1172:"We can still say?" I think what was meant was that, regardless of the dependence of the tests, Ī±ĀÆ ā‰¤ mĪ±percomparison., and the more positively dependent the tests, the more Ī±ĀÆ shrinks toward Ī±. 1180:
That "example" has no explanatory value. What's the point of simply plugging arbitrary values into the formula without providing explanation that connects those values to some useful context?
312:(cf also the Benjamini & Hochberg paper cited at end of the page), so I would be tempted to suggest to move the article to this name, but it may simply be a bias on my side. Any opinionĀ ? 477:
Basically, I first thought it was only a linguistic question when I started this discussion a few months ago. It is only based on the comments above (it was mentioned in particular that
1099:
For example, in fMRI analysis, tests are done on over 100,000 voxels in the brain. The Bonferroni method would require p-values to be smaller than .05/100000 to declare significance.
755:
There is a section in the middle of the article that is repeated word-for-word later in the article. Please fix. Thanx. --Cromwellt 5PM, 16 Feb 2007 (having login problems)
586:
Multiple comparisons. This chapter explains how the comparisons among pairs of the samples tested in an ANOVA test should be done and different test for comparisons like Tukey.
689:
I think that the lead section should be a little more accessable. The big picture in plain language. There is plenty of room for the subtulties of the concept further down.
541:
Disclaimer: I don't know if I am biased by may concrete problem, as I am not statistician, neither I am english native speaker, but I'll try to help. According to dictionary:
489:
the same thing), and the content of the article that I assumed that multiple comparisons (i.e. the ANOVA f-test) was a specific example of multiple testing ā€” while it was not
1072:
In order to retain a prescribed family-wise error rate Ī± in an analysis involving more than one comparison, the error rate for each comparison must be more stringent than Ī±.
267:
Can you add a reference of some kindĀ ? Maybe looks obvious to a statistician, but not so much for me. And I'm a scientistĀ ! Would be very helpful for the general reader.
950: 349:, but I'd be interested to know of any reference that documents the meaning that you indicated above. As I wrote above, I have read mainly papers that use the terminology 786:
may be problematic as they restore claims about Bonferroni that are uncited and have already been called into dispute. Can some of the other editors weigh in on this?
1193:
A marginally less conservative correction can be obtained by solving the equation for the family-wise error rate of m independent comparisons for Ī±percomparison.
713:
I agree as well, and will take a crack at an edit with a more accessible intro, taking into account the current trent in multiple comparisons v testing (above).
547:
Comparison n. 1. The act of comparing; an examination of two or more objects with the view of discovering the resemblances or differences; relative estimate.
1241: 125: 179: 1246: 508:
Yeah, it does seem like we need to rope in some other comments. I'll ask around if anyone has the time to comment on it. Maybe you can do the same?
450:
for the general problem, including in reviews of research papers. Hey, the only paper in the bibliography of the article that mention anything says
758: 353:, but this may be a bias among researchers specialised in a specific area. For example, almost all the literature on the statistical analysis of 528: 367:
Ok, after rereading the article again, it seems to me that it is very confusing in its present stateĀ ! If one reads only the first sentence,
290: 101: 1236: 253: 212: 1213: 951:
http://www.dictionaryofeconomics.com/article?id=pde2010_M000425&edition=current&q=romano%20wolf&topicid=&result_number=1
762: 671: 1081:
Boole's inequality implies that if each of m tests is performed to have type I error rate Ī±/m, the total error rate will not exceed Ī±.
1063:
Multiple testing correction refers to re-calculating probabilities obtained from a statistical test which was repeated multiple times.
837: 274: 593:
In the start of chapter 11: 'The term "multpliple comparisons" was introduced by D. B. Duncan in 1951', according to (David 1995).
1103:
fMRI analysis is often done on over 100,000 voxels, but not always. Also, this is not the best example, since using p-values for
921: 92: 69: 745: 652:'simultaneous inference,' even when the inferences do not deal with comparisons" (from "Multiple hypothesis testing: A review," 957:
We hope Wikipedians on this talk page can take advantage of these comments and improve the quality of the article accordingly.
412:
Sorry, I am lost here; please see the last comment I have written above. It seemed clear to me that this article was about the
605: 162:
This article is or was the subject of a Wiki Education Foundation-supported course assignment. Further details are available
905:
So the question comes down to: what do you want to teach the reader? Currently, the section teaches nothing worthwhile.
44: 1184:
The most conservative method, which is free of dependence and distributional assumptions, is the Bonferroni correction
1107:
in fMRI (as described in the cited sources) isn't quite the same as using p-values in standard null hypothesis tests.
982: 960:
We believe Dr. Wolf has expertise on the topic of this article, since he has published relevant scholarly research:
878: 737:
which I believe is in the public domain, as NIST is a US government agency. In fact I made a template for this:
430:(the general problem). The (good) changes you have made are about the general problem. If the consensus is that 978: 898:
positives and true positives add up to the number of discoveries. Do we need all this quasi-math to know that?
294: 257: 598:
H. A. David First (?) occurrence of common terms in mathematical statistics. Amer. Statist. 49: 121-133, 1995.
216: 1217: 833: 766: 675: 278: 237: 663:, which is authored by and edited by some of the most respected statisticians in the area. In the manual 202:"the likelihood that a fair coin would come up heads at least 9 out of 10 times is 11 * (Ā½)10 = 0.0107." 1168:
If we do not assume that the comparisons are independent, then we can still say: Ī±ĀÆ ā‰¤ m ā‹… Ī±percomparison
998: 870: 50: 829: 375:: if, by multiple comparison, this article means basically "what you do after obtaining a significant 84: 63: 1209: 1012:
I must say it is really ridiculous that something that can be found so easily using a search engine (
917: 913: 909: 825: 431: 270: 208: 32: 974: 851: 735: 446:
record, even though it is not relevant to this particular discussion, I have mostly seen the term
100:
on Knowledge. If you would like to participate, please visit the project page, where you can join
790: 334: 233: 205:
Can someone please explain where the 11 came from? My own understanding is that it should be a 10
187: 623:
merger may be here. I'll plan to do it in a couple of days, if I don't here from anyone else. -
1111:
Since adjacent voxels tend to be highly correlated, this threshold is generally too stringent.
1045: 602: 246:
P(Heads at least 9 out of 10 times)= P(Heads 10 out of 10 times) + P(Heads 9 out of 10 times)
994: 738: 426: 171: 167: 1221: 1049: 1002: 925: 883: 793: 770: 717: 704: 701: 693: 679: 645: 627: 613: 563: 535: 512: 503: 472: 458: 407: 390: 362: 337: 316: 298: 282: 261: 241: 220: 191: 1090:
rate with no multiple comparisons adjustment and the per-test error rates are identical).
464: 17: 807: 624: 354: 1230: 787: 714: 642: 509: 469: 404: 183: 1202:
Continuous generalizations of the Bonferroni and Å idĆ”k correction are presented in.
1041: 690: 742: 610: 560: 500: 455: 387: 359: 313: 158: 329:
to me sound like simultaneous testing of multiple null hypotheses, and that is
865:
to publish a user-friendly guide (and also a workbook) on multiple comparisons
532: 97: 308:
I have always heard of the problems explained in the article under the name
232:
And "likelihood" is the wrong word here; I've changed it to "probability".
1144:
cases, such as bootstrap sampling, they provide only approximate control.
1035: 1030: 1115:
It's POV to say the Bonferroni threshold is "generally too stringent."
1124:
It's POV to refer to other techniques being "better" than Bonferroni.
1025: 1020: 941:, and provided us with the following comments to improve its quality: 569:
Revising some bibliografy, in (Zar, 1999) these are chapters 10 an 11:
416: 379: 813:
shouldn't be "However, for 100 tests where all null hypotheses are
601:
Jerrol H. Zar, Biostatistica Analysis, 4th ed. Prentice-Hall 1999,
799:
Multiple comparisons for confidence intervals and hypothesis tests
413: 376: 731: 399:
I have to agree with Michael Hardy that the term often used is
26: 1055:
Multiple comparisons correction section is highly problematic
544:
Testing n. 1. The act of testing or proving; trial; proof.
889:
Unhelpful section: "Classification of m hypothesis tests"
867:. These sources should be authoritative and mainstream. 577:
other techniques with more than one factor ex. Friedman.
938: 783: 817:, the expected number of incorrect rejections is 5."? 661:
Multiple Testing Problems in Pharmaceutical Statistics
153:
Wiki Education Foundation-supported course assignment
1128:
Such methods can be divided into general categories:
96:, a collaborative effort to improve the coverage of 1176:
Example: 0.2649 = 1 āˆ’ ( 1 āˆ’ .05 ) 6 ā‰¤ .05 Ɨ 6 = 0.3
725: 665:
Multiple Comparisons and Multiple Testing Using SAS
1153:Problems with the "controlling procedures" section 860:. They have also worked with the wizards at the 1085:That sentence is grammatically indecipherable. 946: 8: 386:blur the distinction with multiple testing. 227:It says "at least nine", not "exactly nine". 1014:as I mentioned in the original edit summary 726:Tukey's Studentized Range Test/Distribution 454:, and it is about the general problemĀ ;-). 1207: 304:Multiple comparisons or multiple testingĀ ? 58: 180:Template:Dashboard.wikiedu.org assignment 932: 178:Above undated message substituted from 60: 30: 1206:That's not even a complete sentence. 529:Knowledge talk:WikiProject Mathematics 7: 90:This article is within the scope of 1242:High-importance Statistics articles 730:There is a nice summary of this by 49:It is of interest to the following 933:Dr. Wolf's comment on this article 249:= 1/2^{10} + choose(10,1) 1/2^{10} 25: 157: 110:Knowledge:WikiProject Statistics 83: 62: 31: 1247:WikiProject Statistics articles 494:papers I have seen so far were 130:This article has been rated as 113:Template:WikiProject Statistics 983:22:42, 24 September 2016 (UTC) 317:21:00, 21 September 2005 (UTC) 1: 794:15:38, 13 November 2007 (UTC) 771:23:03, 16 February 2007 (UTC) 628:03:59, 6 September 2006 (UTC) 242:05:57, 31 December 2008 (UTC) 221:05:07, 31 December 2008 (UTC) 104:and see a list of open tasks. 1050:17:43, 19 October 2016 (UTC) 1003:03:18, 18 October 2016 (UTC) 705:07:13, 15 October 2005 (UTC) 391:00:53, 16 October 2005 (UTC) 363:00:03, 16 October 2005 (UTC) 338:22:09, 15 October 2005 (UTC) 333:. So I oppose such a move. 192:04:36, 17 January 2022 (UTC) 1237:C-Class Statistics articles 884:10:13, 24 August 2011 (UTC) 718:08:10, 3 January 2006 (UTC) 694:01:59, 8 October 2005 (UTC) 654:Annual Review of Psychology 646:23:43, 7 January 2007 (UTC) 536:15:08, 5 January 2006 (UTC) 513:05:20, 4 January 2006 (UTC) 504:17:35, 3 January 2006 (UTC) 473:16:54, 3 January 2006 (UTC) 459:15:30, 3 January 2006 (UTC) 408:08:09, 3 January 2006 (UTC) 325:is what I've always heard; 1263: 926:11:36, 22 April 2014 (UTC) 614:11:35, 23 April 2006 (UTC) 564:18:59, 20 April 2006 (UTC) 371:is indeed not the same as 299:14:40, 24 March 2010 (UTC) 283:09:48, 31 March 2009 (UTC) 262:19:14, 26 March 2024 (UTC) 1222:16:49, 27 June 2021 (UTC) 680:17:37, 27 June 2021 (UTC) 345:is exactly what you call 129: 78: 57: 18:Talk:Multiple comparisons 746:18:57, 15 May 2006 (UTC) 954: 937:Dr. Wolf has reviewed 857:Tamhane & Hochberg 465:Bonferroni corrections 441:a set of means, while 93:WikiProject Statistics 39:This article is rated 761:comment was added by 198:Coin flip calculation 166:. Student editor(s): 803:The paragraph says: 638:multiple comparisons 483:multiple comparisons 463:Mathworld says that 432:multiple comparisons 422:multiple comparisons 401:multiple comparisons 384:multiple comparisons 369:multiple comparisons 347:multiple comparisons 323:Multiple comparisons 939:this Knowledge page 902:the table clearer. 852:multiple comparison 116:Statistics articles 331:not the same topic 170:. Peer reviewers: 164:on the course page 45:content assessment 1224: 1212:comment added by 929: 912:comment added by 882: 842: 828:comment added by 774: 424:(this page), and 273:comment added by 211:comment added by 150: 149: 146: 145: 142: 141: 16:(Redirected from 1254: 1008:sources for edit 928: 906: 881: 875: 868: 850:I recommend the 841: 822: 756: 634:Multiple testing 479:multiple testing 452:multiple testing 448:multiple testing 427:multiple testing 373:multiple testing 351:multiple testing 343:multiple testing 327:multiple testing 310:multiple testing 285: 252:= 11 * 1/2^{10} 223: 194: 161: 136:importance scale 118: 117: 114: 111: 108: 87: 80: 79: 74: 66: 59: 42: 36: 35: 27: 21: 1262: 1261: 1257: 1256: 1255: 1253: 1252: 1251: 1227: 1226: 1155: 1136: 1057: 1036:2017 conference 1031:2015 conference 1026:2013 conference 1021:2011 conference 1010: 990: 935: 907: 891: 871: 869: 848: 846:Further reading 823: 808:false positives 801: 780: 757:ā€”The preceding 753: 728: 687: 620:Strong Support. 306: 291:213.173.165.162 268: 206: 200: 177: 155: 132:High-importance 115: 112: 109: 106: 105: 73:Highā€‘importance 72: 43:on Knowledge's 40: 23: 22: 15: 12: 11: 5: 1260: 1258: 1250: 1249: 1244: 1239: 1229: 1228: 1154: 1151: 1131: 1056: 1053: 1039: 1038: 1033: 1028: 1023: 1009: 1006: 989: 986: 975:ExpertIdeasBot 972: 971: 963: 956: 944: 934: 931: 890: 887: 847: 844: 800: 797: 779: 776: 752: 749: 727: 724: 723: 722: 721: 720: 708: 707: 686: 683: 649: 648: 617: 616: 608: 599: 595: 594: 590: 589: 588: 587: 581: 580: 579: 578: 571: 570: 539: 538: 527:Try asking at 524: 523: 522: 521: 520: 519: 518: 517: 516: 515: 397: 396: 395: 394: 393: 355:DNA microarray 305: 302: 265: 264: 254:128.231.234.23 250: 247: 244: 230: 228: 213:67.169.127.132 199: 196: 154: 151: 148: 147: 144: 143: 140: 139: 128: 122: 121: 119: 102:the discussion 88: 76: 75: 67: 55: 54: 48: 37: 24: 14: 13: 10: 9: 6: 4: 3: 2: 1259: 1248: 1245: 1243: 1240: 1238: 1235: 1234: 1232: 1225: 1223: 1219: 1215: 1214:23.242.195.76 1211: 1204: 1203: 1199: 1195: 1194: 1190: 1186: 1185: 1181: 1178: 1177: 1173: 1170: 1169: 1165: 1161: 1160: 1152: 1150: 1149:comparisons? 1146: 1145: 1140: 1135: 1130: 1129: 1125: 1122: 1121: 1116: 1113: 1112: 1108: 1106: 1101: 1100: 1096: 1092: 1091: 1086: 1083: 1082: 1078: 1074: 1073: 1069: 1065: 1064: 1060: 1059:For example: 1054: 1052: 1051: 1047: 1043: 1037: 1034: 1032: 1029: 1027: 1024: 1022: 1019: 1018: 1017: 1015: 1007: 1005: 1004: 1000: 996: 987: 985: 984: 980: 976: 969: 966: 965: 964: 961: 958: 953: 952: 945: 942: 940: 930: 927: 923: 919: 915: 911: 903: 899: 895: 888: 886: 885: 880: 876: 874: 866: 864: 863:SAS Institute 859: 858: 853: 845: 843: 839: 835: 831: 827: 818: 816: 811: 809: 804: 798: 796: 795: 792: 789: 785: 777: 775: 772: 768: 764: 763:67.142.130.42 760: 750: 748: 747: 744: 740: 736: 733: 719: 716: 712: 711: 710: 709: 706: 703: 698: 697: 696: 695: 692: 684: 682: 681: 677: 673: 672:23.242.195.76 668: 666: 662: 657: 655: 647: 644: 639: 635: 632: 631: 630: 629: 626: 625:Scott Alberts 621: 615: 612: 609: 607: 604: 600: 597: 596: 592: 591: 585: 584: 583: 582: 575: 574: 573: 572: 568: 567: 566: 565: 562: 556: 552: 548: 545: 542: 537: 534: 530: 526: 525: 514: 511: 507: 506: 505: 502: 497: 492: 488: 484: 480: 476: 475: 474: 471: 466: 462: 461: 460: 457: 453: 449: 444: 440: 436: 433: 429: 428: 423: 418: 415: 411: 410: 409: 406: 402: 398: 392: 389: 385: 381: 378: 374: 370: 366: 365: 364: 361: 356: 352: 348: 344: 341: 340: 339: 336: 335:Michael Hardy 332: 328: 324: 321: 320: 319: 318: 315: 311: 303: 301: 300: 296: 292: 286: 284: 280: 276: 272: 263: 259: 255: 251: 248: 245: 243: 239: 235: 234:Michael Hardy 231: 229: 226: 225: 224: 222: 218: 214: 210: 203: 197: 195: 193: 189: 185: 181: 175: 173: 169: 165: 160: 152: 137: 133: 127: 124: 123: 120: 103: 99: 95: 94: 89: 86: 82: 81: 77: 71: 68: 65: 61: 56: 52: 46: 38: 34: 29: 28: 19: 1208:ā€”Ā Preceding 1205: 1201: 1200: 1196: 1192: 1191: 1187: 1183: 1182: 1179: 1175: 1174: 1171: 1167: 1166: 1162: 1157: 1156: 1147: 1142: 1141: 1137: 1134: 1127: 1126: 1123: 1118: 1117: 1114: 1110: 1109: 1105:thresholding 1104: 1102: 1098: 1097: 1093: 1088: 1087: 1084: 1080: 1079: 1075: 1071: 1070: 1066: 1062: 1061: 1058: 1040: 1013: 1011: 991: 973: 967: 962: 959: 955: 947: 943: 936: 908:ā€”Ā Preceding 904: 900: 896: 892: 872: 862: 856: 849: 830:Kurai yousei 819: 814: 812: 805: 802: 781: 754: 729: 688: 685:Lead section 669: 664: 660: 658: 653: 650: 637: 633: 619: 618: 557: 553: 549: 546: 543: 540: 495: 490: 486: 482: 478: 451: 447: 442: 438: 434: 425: 421: 400: 383: 372: 368: 350: 346: 342: 330: 326: 322: 309: 307: 287: 275:146.50.10.49 266: 204: 201: 176: 156: 131: 91: 51:WikiProjects 995:Jbrowning17 824:ā€”Preceding 269:ā€”Preceding 207:ā€”Preceding 172:Jbrowning17 168:Jbrowning17 1231:Categories 1159:increases. 1120:negatives. 968:Reference 914:AndrewOram 784:reversions 751:Redundancy 606:013081542X 107:Statistics 98:statistics 70:Statistics 988:Criticism 879:Wolfowitz 439:comparing 1210:unsigned 1077:topics. 922:contribs 910:unsigned 855:book by 838:contribs 826:unsigned 759:unsigned 715:Debivort 643:Septagon 510:Debivort 470:Debivort 405:Debivort 271:unsigned 209:unsigned 184:PrimeBOT 1042:Orielno 782:Recent 739:NIST-PD 691:ike9898 443:testing 134:on the 41:C-class 873:Kiefer 791:Bivort 743:Btyner 611:Arauzo 561:Arauzo 501:Schutz 456:Schutz 417:F-test 388:Schutz 380:F-test 360:Schutz 314:Schutz 47:scale. 533:linas 485:were 414:ANOVA 377:ANOVA 1218:talk 1046:talk 999:talk 979:talk 918:talk 834:talk 815:true 767:talk 732:NIST 702:Jake 676:talk 636:and 603:ISBN 481:and 295:talk 279:talk 258:talk 238:talk 217:talk 188:talk 126:High 810:." 778:RVs 734:at 496:all 487:not 182:by 1233:: 1220:) 1048:) 1001:) 981:) 924:) 920:ā€¢ 840:) 836:ā€¢ 788:de 769:) 741:. 678:) 531:. 491:my 435:is 297:) 281:) 260:) 240:) 219:) 190:) 174:. 1216:( 1044:( 997:( 977:( 916:( 877:. 832:( 773:. 765:( 674:( 293:( 277:( 256:( 236:( 215:( 186:( 138:. 53:: 20:)

Index

Talk:Multiple comparisons

content assessment
WikiProjects
WikiProject icon
Statistics
WikiProject icon
WikiProject Statistics
statistics
the discussion
High
importance scale

on the course page
Jbrowning17
Jbrowning17
Template:Dashboard.wikiedu.org assignment
PrimeBOT
talk
04:36, 17 January 2022 (UTC)
unsigned
67.169.127.132
talk
05:07, 31 December 2008 (UTC)
Michael Hardy
talk
05:57, 31 December 2008 (UTC)
128.231.234.23
talk
19:14, 26 March 2024 (UTC)

Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.

ā†‘