Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Bayesian Tests to Quantify the Result of a Replication Attempt Josine Verhagen & Eric-Jan Wagenmakers University of Amsterdam Abstract Replication attempts are essential to the empirical sciences. Successful replication attempts increase researchers’ confidence in the presence of an effect, whereas failed replication attempts induce skepticism and doubt. However, it is often unclear to what extent a replication attempt results in success or failure. To quantify replication outcomes we propose a novel Bayesian replication test that compares the adequacy of two competing hypotheses. The first hypothesis is that of the skeptic and holds that the effect is spurious; this is the null hypothesis that postulates a zero effect size, H0 : δ = 0. The second hypothesis is that of the proponent and holds that the effect is consistent with the one found in the original study, an effect that can be quantified by a posterior distribution. Hence, the second hypothesis –the replication hypothesis– is given by Hr : δ ∼ “posterior distribution from original study”. The weighted likelihood ratio between H0 and Hr quantifies the evidence that the data provide for replication success and failure. In addition to the new test, we present several other Bayesian tests that address different but related questions concerning a replication study. These tests pertain to the independent conclusions of the separate experiments, the difference in effect size between the original experiment and the replication attempt, and the overall conclusion based on the pooled results. Together, this suite of Bayesian tests allows a relatively complete formalization of the way in which the result of a replication attempt alters our knowledge of the phenomenon at hand. The use of all Bayesian replication tests is illustrated with three examples from the literature. For experiments analyzed using the t test, computation of the new replication test only requires the t values and the numbers of participants from the original study and the replication study. Keywords: Effect Size, Prior Distribution, Bayes Factor. QUANTIFYING REPLICATION RESULTS 2 Introduction In recent years, the field of psychology has come under increased scrutiny. The prominence of implausible results, research fraud, questionable research practices, lack of data sharing, confirmation bias, publication bias, and failure to replicate key findings have led some researchers to conclude that psychology is in a “crisis of confidence” (e.g., Pashler & Wagenmakers, 2012; see also Bones, 2012; Francis, in pressb; John, Loewenstein, & Prelec, 2012; Neuroskeptic, 2012; Simmons, Nelson, & Simonsohn, 2011; Wagenmakers, Wetzels, Borsboom, & van der Maas, 2011; Wicherts, Borsboom, Kats, & Molenaar, 2006). One of the proposed solutions to overcome this crisis is to re-emphasize the importance of independent replications (e.g., Asendorpf et al., 2013; LeBel & Peters, 2011; Nosek & Bar–Anan, 2012; Nosek, Spies, & Motyl, 2012; Yong, 2012). The increased emphasis on replications is evident from initiatives such as psychfiledrawer.org, the new replication sections in Perspectives on Psychological Science and Attention, Perception, & Psychophysics, the special issues on replications in Social Psychology (Nosek & Lakens, 2013) and Frontiers in Cognition (Zwaan & Zeelenberg, 2013), and the Open Science Framework project on reproducibility.1 Spearheaded by Brian Nosek, the latter initiative combines the research efforts of over 80 researchers to try and replicate a selection of articles from the 2008 volumes of Journal of Experimental Psychology: Learning, Memory, & Cognition, Psychological Science, and the Journal of Personality and Social Psychology. Now that the field is on the verge of a paradigmatic shift towards increased transparency and more replication attempts, new questions arise. Most acute is the question of how to judge and quantify replication success. Qualitatively, many aspects need to be considered when it comes to deciding whether a replication attempt resulted in success or failure (Brandt et al., 2013). Did the researchers test a similar population of participants? Did the researchers alter the paradigm in a meaningful way? Here we ignore these qualitative judgments and focus instead on the main quantitative aspect: assuming the experimental design and data collection strategy are adequate, how does one analyze the data to measure the statistical evidence for success or failure of a replication attempt? In classical or frequentist statistics, the success of a replication attempt is commonly decided upon by comparing the p values of the original experiment and the replication attempt. However, comparing p values is a practice that is fraught with problems, as a difference in significance does not always indicate that the difference is significant (Gelman & Stern, 2006; Nieuwenhuis, Forstmann, & Wagenmakers, 2011) and differences in power between the original and replicated study can lead to counter-intuitive conclusions. An alternative procedure is to compare the effect size estimates in the two studies directly, either by computing a p value or by inspecting confidence intervals (Cohen, 1988, 1990; Rosenthal & Rubin, 1990). One problem with this procedure is that it can be difficult to demonstrate a failure to replicate when the original experiment was so low in power that 1 See http://www.openscienceframework.org/project/EZcUj/wiki/home. This research was supported by an ERC grant from the European Research Council. Correspondence concerning this article may be addressed to Josine Verhagen, University of Amsterdam, Department of Psychology, Weesperplein 4, 1018 XA Amsterdam, the Netherlands. Email address: A.J.Verhagen@uva.nl. QUANTIFYING REPLICATION RESULTS 3 the confidence interval on effect size is relatively broad and almost covers zero (Asendorpf et al., 2013). To overcome these and other limitations, Simonsohn (2013) recently proposed a twostep method. In the first step, one defines what constitutes a small effect (in terms of how detectable it is given the sample size in the original study). In the second step, one tests whether the effect in the replication attempt is smaller than this small effect; when the effect is smaller than small, one concludes that the replication attempt was not successful. One disadvantage of this method –and frequentist methods in general– is that it cannot quantify support in favor of the null hypothesis. Thus, if a replication attempt is indeed successful (or, in frequentist parlance, not unsuccessful) the Simonsohn procedure cannot easily quantify the strength of evidence that the data provide for this success. A third way in which to analyze the results of a replication attempt is to conduct a meta-analysis and assume that the original study and the replication attempt are exchangeable (Rouder & Morey, 2011). To assess whether or not there is a true effect, a common measure of effect size is obtained by weighting the effect sizes in the respective studies by the precision with which they were measured. In general, meta-analyses are based either on a fixed-effect model or on a random-effects model. The fixed-effect model assumes that all factors that could influence the effect size are the same in all studies. Hence, variations in effect size are due solely to sampling error around the true effect size. In contrast, the random-effects model assumes that each study has its own factors influencing effect size, and therefore the effect sizes are similar but not identical across studies. Effect size differences between studies are therefore caused both by sampling error and by between-study variance, where the effect size for each individual study is assumed to be normally distributed around a mean (overall) effect size. The drawback of this latter approach is that between-study variance is difficult to estimate when the number of available studies is limited. This means that in case of a single replication attempt, the random-effects approach is not feasible. In Bayesian statistics, the success of a replication attempt has been decided upon using procedures that mimic those in the frequentist toolkit. For instance, based on the replication attempt one can quantify evidence for and against the null hypothesis using a default Bayesian hypothesis test (Dienes, 2011; Jeffreys, 1961; Kass & Raftery, 1995; Lee & Wagenmakers, 2013). This test answers the question “Is the effect present or absent in the data from the replication attempt?”. When the test outcome supports the null hypothesis, this can be taken as evidence for replication failure (e.g., Shanks et al., 2013; Ullrich, Krueger, Brod, & Groschupf, in press; Wagenmakers, Wetzels, Borsboom, van der Maas, & Kievit, 2012). Despite its intuitive appeal, this default Bayes factor hypothesis test often does not relate the data of the replication attempt to the data from the original experiment (but see Dienes, 2011). Consequently, the comparison of Bayes factors and p values share the same limitation: when the original experiment yields support in favor of the alternative hypothesis, and the replication attempt yields support in favor of the null hypothesis, this does not necessarily signal support for the assertion that the replication attempt differs from the original experiment. To overcome this limitation, Bayarri and Mayoral (2002b) developed a Bayesian hypothesis test to assess whether the difference in effect size between the original experiment and the replication attempt is equal to zero (i.e., H0 ) or not (i.e., H1 ). This test answers the question “Are the effects from both studies of the same size?”. When the test outcome sup- QUANTIFYING REPLICATION RESULTS 4 ports the null hypothesis, this can be taken as evidence for exact replication. The limitation of this procedure –which it shares with frequentist comparisons of effect size– is that it can be difficult to find evidence for either hypothesis in case of a low-powered experiment. The example used in Bayarri and Mayoral (2002b) makes this clear: with 10 participants per condition, even an exact replication does not yield a test result that supports the replication hypothesis in a compelling fashion. Finally, the fixed-effect Bayesian meta-analysis test (Rouder & Morey, 2011) assesses the extent to which the combined data from all pertinent studies provides evidence for or against the null hypothesis of no effect. This test answers the question “When we pool the information from all experiments, do the data support the absence or the presence of the effect?”. This Bayesian test assumes that all experiments are exchangeable. Specifically, it assumes that each experiment was correctly executed in a purely confirmatory setting (De Groot, 1956/in press; Wagenmakers et al., 2012), as an exact replication of the original study; it also assumes that each experiment has the exact same underlying effect size. Each of these assumptions may be questioned, particularly by skeptics who have grounds to doubt the reproducibility of the findings from the original experiment. Here we propose a new Bayesian test for replication success. Inspired by the work of Simonsohn (2013) and Dienes (2011), we seek to quantify the success or failure of a replication attempt. Given that the original experiment reported a significant effect, this test addresses the question “Is the effect from the replication attempt comparable to what was found before, or is it absent?”. Our test is based on a straightforward adaptation of the standard Bayesian hypothesis test, using the data from the replication attempt to pit against one another the following two hypotheses: first, the null hypothesis H0 – this represents the idealized belief of a skeptic, who has reason to doubt the presence of a particular effect. For the null hypothesis, effect size equals zero, that is, H0 : δ = 0.2 The second hypothesis is the replication hypothesis Hr – this represents the idealized belief of a proponent, often the person who carried out the original experiment but more generally any researcher who feels that the results of the original experiment(s) can be trusted.3 Here is where our test deviates from the default Bayesian hypothesis tests advocated in psychology by Rouder, Morey, and ourselves (Rouder, Speckman, Sun, Morey, & Iverson, 2009; Rouder, Morey, Speckman, & Province, 2012; Wagenmakers, van der Maas, & Grasman, 2007; Wetzels & Wagenmakers, 2012; Wetzels, Grasman, & Wagenmakers, 2012). In the default Bayesian hypothesis tests, H0 is compared against the alternative hypothesis H1 that features a relatively vague belief about effect size, often quantified by a Gaussian distribution centered on zero. This default assignment of belief is blind to the earlier experiments that have been observed (Dienes, 2011). In contrast, the idealized belief of the proponent is tightly linked to the outcome of the original experiment. As outlined below, we assume that the idealized belief of the proponent is given by the posterior distribution on effect size as obtained from the original experiment. Hence, our replication test assesses the extent to which the data from the replication attempt are more likely under the skeptic’s hypothesis H0 : δ = 0 or the proponent’s replication hypothesis 2 It is possible to relax this assumption and also allow small values of δ surrounding zero (Morey & Rouder, 2011), but we will not pursue such an analysis here. 3 The general idea of specifying the alternative hypothesis based on results from earlier studies was also put forward and implemented by Dienes (2011) and Bem, Utts, and Johnson (2011). QUANTIFYING REPLICATION RESULTS 5 Hr : δ ∼ “posterior distribution from original study”. The default test and our replication test address important yet complementary questions regarding the interpretation of data from replication attempts. To reiterate, the default test addresses the question “Given that we know relatively little about the expected effect size beforehand, is the effect present or absent in the replication attempt?”, and our test addresses the question “Is the effect similar to what was found before, or is it absent?”. The two tests therefore represent extremes on a continuum of sensitivity to past research; the default test completely ignores the outcomes of an earlier experiment, whereas the replication test takes these outcomes fully into account. The new test also deviates from the test for equality of effect sizes (Bayarri & Mayoral, 2002b) because the latter test does not take into account the skeptic’s hypothesis that the effect is absent. Finally, the new test deviates from the fixed-effect meta-analysis test (Rouder & Morey, 2011), because, from a skeptic’s position, the replication attempt is not exchangeable with the original experiment. For concreteness and simplicity, the remainder of this article focuses on experiments analyzed using the t test. We first describe the existing Bayesian tests as they apply to replication attempts. Subsequently, we discuss the new Bayesian replication test in more detail and then explain how the posterior distribution can be derived from the t test results of the original experiment. Next we present a simulation study to demonstrate that the test works as intended and compare its results to those obtained from existing Bayesian tests. Finally, we apply each Bayesian test to three example findings from the published literature: the finding that women prefer men in red (Elliot et al., 2010), the finding that priming participants with concepts related to intelligence makes them act more intelligently (Dijksterhuis & van Knippenberg, 1998; LeBoeuf & Estes, 2004; Shanks et al., 2013), and the finding that negative priming is observed even when the prime was unattended (Milliken, Joordens, Merikle, & Seiffert, 1998; Neill & Kahan, 1999). Bayes Factor Hypothesis Tests Consider a standard one-sample t test for the difference between two dependent measures Y1 and Y2 . Assume that both scores are normally distributed and √ that the means and variances are unknown. In this case, the quantity t = (ȳ1 − ȳ2 )/(s/ N ) has a Student-t distribution with df = N − 1 degrees of freedom, where ȳi is the mean score on measure Yi , s is the standard deviation of the difference between √ Y1 and Y2 , and N is the number of observations. The effect size δ equals (ȳ1 − ȳ2 )/s = t/ N .4 Under the null hypothesis H0 : δ = 0, the distribution of t is symmetric and centered at zero. Under the hypothesis of a specific alternative effect√size H1 : δ = δ1 , the distribution of t is non-central with non-centrality parameter ∆1 = δ1 N . The models H0 and H1 are each defined by a single specific effect size. Hence is it straightforward to compute, for each model, the probability of the observed data under the hypothesized effect size. The ratio of these probabilities is a likelihood ratio (Royall, 1997), and it quantifies the extent to which 4 a two sample t test, for equal variances replace s with sp = p For ((n1 − 1)s21 + (n2 − 1)s22 )/(n1 + n2 − 2), N with Np = 1/(1/n1 + 1/n2 ) and df with dfpp= n1 + n2 − 2, √ where ni is the number of participants in group i. For unequal variances replace s N with s21 /n1 + s22 /n2 2 2 2 2 2 2 2 and df with (s1 /n1 + s2 /n2 ) /((s1 /n1 ) /(n1 − 1) + (s2 /n2 ) /(n2 − 1)). QUANTIFYING REPLICATION RESULTS 6 the observed data Y are more likely to have occurred under H1 relative to H0 : LR10 = tdf, δ1 √N (tobs ) p(Y | H1 ) = , p(Y | H0 ) tdf (tobs ) (1) where tdf, δ1 √N (tobs ) is the ordinate of the non-central t-distribution and tdf (tobs ) is the ordinate of the central t-distribution, both with df degrees of freedom and evaluated at the observed value tobs . The problem with this analysis is that the exact alternative effect size δ1 is never known beforehand. In Bayesian statistics, this uncertainty about δ is addressed by assigning it a prior distribution. Prior distributions reflect the relative plausibility of the entire range of different effect sizes. For instance, in the absence of subject-specific knowledge about the phenomenon of interest, a default prior distribution is H1 : δ ∼ N (0, 1), a standard normal distribution (Rouder et al., 2009): this distribution reflects the belief that effect size is likely to be close to zero, and any positive value is just as likely as its corresponding negative value. Other prior distributions are possible and they will yield different results. After assigning δ a prior distribution, the immediate computation of the likelihood ratio (Equation 1) is frustrated by the fact that H1 is now a composite hypothesis, reflecting an entire distribution of effect sizes instead of a single value. Mathematically, this complication is overcome by computing a weighted average of likelihood ratios across all possible values of effect size, where p(δ | H1 ) serves as the weight function. To compute this weighted average, an integral is taken with respect to effect size. The resultant average likelihood ratio is better known as the Bayes factor: p(Y | H1 ) p(Y | H0 ) R p(Y | δ, H1 )p(δ | H1 ) dδ = p(Y | H0 ) Z t √ (t )p(δ | H ) dδ 1 df, δ N obs = tdf (tobs ) Z = LR10 p(δ | H1 ) dδ. B10 = (2) Interpretation of the Bayes factor The Bayes factor B10 quantifies the strength of the evidence that the data provide for H1 versus H0 . When B10 = 11.5, for instance, the data are 11.5 times more likely to have occurred under H1 than under H0 . Values of B10 lower than 1 support H0 . In such cases, a better appreciation for the strength of evidence in favor of H0 can be obtained by computing B01 = 1/B10 (Rouder & Morey, 2012); for instance, when B10 = .04 this means that the data from the replication attempt are 1/.04 = 25 times more likely to have occurred under H0 than under H1 . The outcome of a Bayes factor test is not a decision. Instead, “its function is to grade the decisiveness of the evidence.” (Jeffreys, 1961, p. 432). Although the decisiveness of evidence has a continuous scale, Jeffreys (1961) proposed a series of discrete categories that QUANTIFYING REPLICATION RESULTS 7 facilitate scientific communication (see also Wetzels et al., 2011). In particular, Jeffreys (1961) argued that values of B10 in between 3 and 1/3 are anecdotal or “not worth more than a bare mention”. This label is helpful because it explicitly allows for the possibility that the data are ambiguous, and because it protects researchers from over-interpreting their findings. Note that, in contrast to p values, Bayes factors are able to quantify evidence in favor of a null hypothesis H0 . This is particularly relevant for replication research. Three Extant Bayes Factors for the Analysis of Replication Attempts This section details three extant Bayes factor tests that can be used to analyze the results of a replication attempt. This will set the stage for the introduction of our new test. As discussed above, these Bayes factors address different but related questions. Specifically, the default Bayes factor test addresses the question “Is the effect present or absent in the replication attempt?”, the Bayes factor test for equality of effect sizes addresses the question “Are the effect sizes from the original experiment and the replication attempt the same?”, and the fixed-effect meta-analysis test addresses the question “When we treat all pertinent studies as exchangeable, is the effect present overall?”. 1. Independent JZS Bayes Factor Test: “Is the Effect Present or Absent in the Replication Attempt?” The default Bayes factor test assesses the evidence for the presence or absence of an effect in isolation, without recourse to the original experiment. The test compares the null hypothesis that effect size δ is zero against an alternative hypothesis that effect size is not zero. The alternative hypothesis, however, needs to be specified explicitly and precisely. Because the alternative hypothesis is so well-specified, the Bayes factor can quantify the extent to which the data support that hypothesis versus the null hypothesis. Consequently, Bayes factors have two concrete advantages over classical null hypothesis significance testing by means of p-values. The first advantage is that Bayes factors can quantify evidence in favor of the null hypothesis (Rouder et al., 2009), something that is of primary importance in hypothesis tests for replication attempts; the second advantage is that Bayes factors are relatively reluctant to reject the null hypothesis when it provides a poor fit – after all, the alternative hypothesis may provide a fit that is just as poor or even worse (Edwards, Lindman, & Savage, 1963; Sellke, Bayarri, & Berger, 2001). To specify the alternative hypothesis that the effect size is different from zero (i.e., H1 : δ 6= 0), the default test assigns a general-purpose prior to effect size. Most default general-purpose priors are centered on zero and have fat tails (for details see Bayarri, Berger, Forte, & Garcı́a-Donato, 2012). For the t test, a popular prior is the standard Cauchy prior δ ∼ Cauchy(0, 1) (Rouder et al., 2009). As can be seen in Figure 1, the Cauchy distribution has fatter tails than the standard Normal distribution, and assigns highest probability to effect sizes between −2 and 2, which contains most effect sizes in psychology. Together with uninformative priors on the means and variances, this popular prior set-up is also known as the Jeffreys-Zellner-Siow (JZS) prior (Liang, Paulo, Molina, Clyde, & Berger, 2008; Zellner, 1986; Zellner & Siow, 1980; Bayarri & Garcı́a-Donato, 2007; Bayarri et al., 2012). After explicitly specifying the alternative hypothesis, the JZS Bayes factor is now computed as the ratio of two probabilities: the probability of the observed data given the null QUANTIFYING REPLICATION RESULTS 8 Figure 1. The Cauchy distribution and the standard Normal distribution. hypothesis H0 : δ = 0, and the probability of the observed data given H1 : δ ∼ Cauchy(0, 1). As explained in the previous section, the latter probability is obtained by averaging the likelihood of the observed data with respect to the Cauchy prior distribution. The resulting Bayes factor, B10 , quantifies the extent to which the observed data support H1 over H0 : p(Y | H1 ) p(Y | H0 ) R p(tobs | δ)p(δ | H1 ) dδ = p(tobs | δ = 0) R tdf,δ√N (tobs )p(δ | H1 ) dδ = , tdf (tobs ) B10 = (3) p(δ | H1 ) ∼ Cauchy (0, 1) As indicated in the previous section, values of B10 higher than 1 indicate support for H1 , whereas values lower than 1 indicate support for H0 (i.e., something that is not possible with p-values). 2. Equality-of-Effect-Size Bayes Factor Test: “Does the Effect Size in the Replication Attempt Equal the Effect Size in the Original Study?” Both proponents and skeptics should be interested in the assessment of whether the effect size from an original experiment equals that of the replication attempt. Bayarri and Mayoral (2002b) proposed a Bayes factor test to quantify the evidence for the hypothesis that the effect sizes from two experiments are equal versus the hypothesis that they are not. QUANTIFYING REPLICATION RESULTS 9 In their test, Bayarri and Mayoral (2002b) assumed that there was one true underlying effect size µ, from which the effect sizes in the original (δorig ) and replicated study (δrep ) deviate with variance τ 2 . The test for the effect size difference between the original experiment and the replication attempt therefore amounts to testing whether this variance τ 2 is equal to zero: p(Yorig , Yrep | H0 ) B01 = p(Yorig , Yrep | H1 ) (4) p(Yorig , Yrep | δorig , δrep )p(δorig , δrep | τ 2 = 0) R = . p(Yorig , Yrep | δorig , δrep )p(δorig , δrep | τ 2 )p(τ 2 | H1 ) dτ 2 The statistical details of this equality-of-effect-size Bayes factor are reported in Appendix B. Note that for this particular Bayes factor, the null hypothesis corresponds to a lack of difference between the effect sizes. Hence, evidence in favor of the null hypothesis is indicative of a successful replication. For consistency of interpretation with the other Bayes factors, for which the null hypothesis represents replication failure instead of success, the remainder of this paper reports this Bayes factor as B01 instead of as B10 . 3. Fixed-Effect Meta-Analysis Bayes Factor Test: “When Pooling All Data, is the Effect Present or Absent?” When it is plausible to treat a series of experiments as exchangeable, and when there is no scepticism about the original experiment, one may wonder about the extent to which the pooled data support the hypothesis that a particular effect is present or absent. This question can be addressed by a fixed-effect meta-analytic Bayes factor (Rouder & Morey, 2011). The test assumes that there is a single true underlying effect size, and that the observed effect sizes in the various experiments fluctuate around this true value due only to sampling variability. The JZS Bayes factor described earlier can be generalized to include M independent experiments. As before, a Cauchy distribution is taken as a prior distribution for δ in each experiment. As the experiments are independent, this results in a product of the terms in Equation 3 over all M studies: p(Y1 , . . . , YM | H1 ) p(Y , . . . , YM | H0 ) R 1 p(t1 , . . . , tM | δ)p(δ | H1 ) dδ = p(t1 , . . . , tM | δ = 0) i R hQ M m=1 p(tm | δ) p(δ | H1 ) dδ = QM m=1 p(tm | δ = 0) i R hQ M √ m=1 tdfm ,δm N (tm ) p(δ | H1 ) dδ , = QM t (t ) m df m=1 B10 = (5) p(δ | H1 ) ∼ Cauchy (0, 1) . The resulting meta-analytic Bayes factor quantifies the evidence that the data provide for the hypothesis that the true effect size is present (i.e., H1 ) versus absent (i.e., H0 ). Hence, a high Bayes factor B10 indicates that the evidence from the pooled data support the hypothesis that a true effect is present. QUANTIFYING REPLICATION RESULTS 10 The New Bayes Factor Test for Replication Success In this section we outline the rationale of our novel Bayesian replication test as it applies to the common Student t test. The central idea is that we can pit against each other two hypotheses for data Yrep from a replication attempt: the skeptic’s null hypothesis H0 and the proponent’s replication hypothesis Hr . The relative support of data Yrep for H0 versus Hr is quantified by a Bayes factor, whose computation requires that we specify Hr precisely. As explained earlier, the new replication test defines Hr to be the idealized belief of a proponent, that is, the posterior distribution from the original experiment, p(δ | Yorig ). Given this definition of Hr , a straightforward way to approximate the desired Bayes factor Br0 is to draw M samples from p(δ | Yorig ), compute a likelihood ratio for each sample, and then average the likelihood ratios: p(Yrep | Hr ) p(Yrep | H0 ) R p(Yrep | δ, Hr )p(δ | Yorig ) dδ = p(Yrep | H0 ) R √ tdf,δ N (trep )p(δ | Yorig ) dδ = tdf (trep ) Br0 = ≈ M 1 X tdf,δ(i) √N (trep ) , M tdf (trep ) i=1 (6) δ (i) ∼ p(δ | Yorig ), where trep denotes the observed t value in the replication attempt, and the approximation in the last step can be made arbitrarily close by increasing the number of samples M . As detailed in a later section, only the t value torig and the degrees of freedom from the original experiment are used to determine p(δ | Yorig ); together with the summary data from the replication attempt (i.e., trep and df ) this provides all information required to obtain Br0 as indicated in the final step from Equation 6. R code to calculate Br0 is available on the first author’s website.5 A Frequentist Replication Factor Our new test is Bayesian but it is possible to develop frequentist analogues. One idea is to compute a likelihood ratio, as in Equation 1, and plug in the single best estimate of effect size as an instantiation of the replication hypothesis: δ = δ̂ | Yorig . A major drawback of this procedure is that it is based on a point estimate, thereby ignoring the precision with which the effect size is estimated – and precisely estimated effect sizes lead to more informative tests. A more sophisticated idea is to bootstrap the original data set ⋆,i (Efron & Tibshirani, 1993) and obtain, for each bootstrapped data set Yorig , a corresponding ⋆,i best estimate of effect size, δ̂ ⋆,i | Yorig . The distribution of these bootstrapped effect size estimates can be viewed as a poor man’s Bayesian posterior distribution. Next, a frequentist ⋆,i counterpart of Equation 6 can be obtained by computing a likelihood ratio LR0r for every ⋆ ⋆ one of K bootstrapped effect size estimates δ̂ | Yorig , and then averaging these likelihood P ⋆,i ratios to obtain a “frequentist replication factor”: F RF0r ≈ K1 K i=1 LR0r . 5 http://www.josineverhagen.com/?page_id=76. QUANTIFYING REPLICATION RESULTS 11 From Prior to Posterior to Prior to Posterior: Bayesian Updating So far our analysis assumed that we have available the posterior distribution of the effect size obtained from the original study, p(δ | Yorig ). This posterior is a compromise between the prior distribution and the information in the data, and we deal with these two ingredients in turn. First, for the prior, we assume that before seeing the data from the original study the proponent started out with an uninformative, flat prior distribution on effect size. This is an assumption of convenience that is, fortunately, not at all critical: under general conditions, the data overwhelm the prior and dissimilar prior distributions result in highly similar posterior distributions (Edwards et al., 1963). The flat distribution is convenient to use because it allows a connection to the frequentist literature on confidence intervals. Second, for the information in the data from the original experiment, we only need the t value and the degrees of freedom. Under the assumption of a flat prior, we can then obtain the posterior distribution for effect size. This posterior distribution√is a mixture of noncentral tdf,∆ (tobs )-distributions with non-centrality parameters ∆ = δ N . To approximate this posterior distribution, we first use an iterative algorithmic search procedure (Cumming & Finch, 2001) to obtain a 95% frequentist confidence interval for effect size; next we interpret this 95% confidence interval as a 95% Bayesian credible interval (Lindley, 1965; Rouanet, 1996) and assume the posterior is normally distributed. As demonstrated in Appendix A, the normal approximation to the posterior is highly accurate, even for small sample sizes. For the proponent then, the Bayesian updating sequence proceeds as follows: before observing the data from the original experiment, the proponent has an uninformative prior distribution on effect size, p(δ); when the data from the original experiment become available, this prior distribution is updated to a posterior distribution, p(δ | Yorig ); this posterior reflects the proponent’s idealized belief about δ after observing the original experiment, and it is used as a prior for our Bayesian replication test, as per Equation 6. Finally, the posterior-turned-prior p(δ | Yorig ) is confronted with the data from the replication attempt and is updated to another posterior, p(δ | Yrep , Yorig ).6 Simulation Study To illustrate the behavior of the different Bayesian tests, Tables 1-4 show the results of a simulation study based on a one-sample t test. The study systematically varied the t values and sample sizes for an original experiment and a replication attempt, resulting in different replication test results. The rows are defined by the t values in the original study and feature original sample sizes of 50 (upper half) and 100 (lower half). The columns are defined by the t values in the replication attempt and feature replication sample sizes of 50 (left half) and 100 (right half). 6 From the point of view of the skeptic, the updating process ignores the possibility that the effect is absent; in addition, there may be reasons to doubt that Yorig is tainted by publication bias or hindsight bias. Our replication test quantifies the extent to which the replication data Yrep support the proponent’s beliefs quantified by p(δ | Yorig ) versus the skeptic’s beliefs that are quantified by δ = 0. QUANTIFYING REPLICATION RESULTS 12 New Bayes Factor Replication Test For the Bayes factor replication test, the simulation study highlights three intuitive regularities. First, evidence for successful replication increases when the replicated effect size is further away from the skeptic’s belief that the effect size is zero. In Table 1, each separate row shows that Br0 increases with δrep . Second, evidence for successful replication increases when the replicated effect size δrep is closer to the observed effect size δorig . For example, within the (norig = nrep = 50) quadrant, consider the column with a replicated effect size of .28. Within this column, the Bayes factor is highest (i.e., 4.94) when the original effect size is also .28 (i.e., the second row). The next column, with a replicated effect size of .42, shows a similar effect: again, the Bayes factor is highest (i.e., 46) when the original effect size is also .42 (i.e., the third row). This regularity is apparent in all quadrants of Table 1. Third, more decisive test outcomes are obtained when the sample size in the replication attempt is large. With increased sample size, there is more support for the null hypothesis in case the replication effect size is close to zero, and there is more support for the replication hypothesis in case the replication effect size equals or exceeds the original effect size. For example, the lower left panel in Table 1 shows that Br0 = 5.67 with norig = nrep = 50, δorig = .3, and δorig = .28; the lower right panel shows that for comparable effect sizes but twice as many participants, a more informative outcome is obtained: Br0 = 53.69. A similar but much weaker regularity is apparent when the sample size of the original experiment is increased. These Bayes factor results are similar to those obtained from the method proposed by Dienes (2011), who used a normal prior distribution with mean Morig and standard deviation Morig /2, where Morig is the mean observed in the original experiment. Although the results of the Bayes factor replication test are on a continuous scale we can compare them to the common practice of labeling replications as success or failure based on the p value from the replicated study. According to a conventional two-sided t test with 50 or 100 participants and an alpha level of .05, the null hypothesis should be rejected at a t value near 2. In Table 1, the results for the columns where trep = 2 demonstrate that this corresponds to the data being at worst 0.53 and at best 5.67 times more likely to occur under the replication hypothesis than under the null hypothesis. Across most entries in Table 1, a replication t value of 2 yields evidence that favors the replication hypothesis, albeit in rather modest fashion (Jeffreys, 1961). This illustrates the well-known fact that the p value overestimates the evidence against the null hypothesis (Edwards et al., 1963; Sellke et al., 2001). The overestimation occurs because the p value does not take into account the adequacy of the alternative hypothesis; when both the null hypothesis and the alternative hypothesis provide an equally poor account of the data, the p value indicates that the null hypothesis should be rejected. For the same data, the Bayesian replication test will correctly indicate that the data are ambiguous. Independent JZS Bayes Factor Test Table 2 shows the outcomes for the JZS Bayes factor test. For the evaluation of the results from the replication attempt, the JZS Bayes factor test ignores the findings that were obtained in the original study. Hence, within each of the four quadrants all rows in QUANTIFYING REPLICATION RESULTS Replication torig δorig trep δrep 1 0.14 2 0.28 3 0.42 4 0.57 1 2 3 4 trep δrep 0.1 0.2 0.3 0.4 Bayes Factor Br0 norig = 50, nrep = 50 0 1 2 3 4 .00 .14 .28 .42 .57 0.55 1.17 3.87 17.71 96 0.26 0.91 4.94 36.50 311 0.08 0.44 3.88 46.01 620 0.02 0.14 1.94 36.49 770 norig = 100, nrep = 50 0.00 0.71 1.41 2.12 2.83 .00 .1 .2 .3 .4 0.69 1.05 1.86 3.75 8.24 0.42 0.89 2.20 6.11 18.29 0.19 0.54 1.87 7.19 29.45 0.06 0.24 1.15 6.13 34.51 0.00 .00 0.41 0.15 0.03 0.00 0 .00 0.55 0.26 0.08 0.01 13 norig = 50, nrep = 100 0.71 1.41 2.12 2.83 .14 .28 .42 .57 1.53 18.82 585 44343 1.09 25.64 1513 222076 0.42 18.28 2019 576871 0.09 7.28 1449 783072 norig = 100, nrep = 100 1 2 3 4 .1 .2 .3 .4 1.17 3.97 20.28 141 0.91 5.08 42.25 472 0.44 3.97 53.68 966 0.13 1.93 42.24 1220 Table 1: Results from the Bayes factor replication test Br0 applied to different values of torig , trep , norig , and nrep . The rows are defined by the t values in the original study and corresponding effect sizes for original sample sizes of 50 (upper half) and 100 (lower half). The columns are defined by the t values in the replication attempt and corresponding effect sizes for replication sample sizes of 50 (left half) and 100 (right half). Table 2 are identical. The JZS Bayes factors show increasing evidence for the null over the alternative hypothesis when the effect size in the replicated study increases. This increase is stronger for larger sample sizes. The results show that if the effect size of the replication attempt is in the same direction as that of the original experiment, the default test generally yields weaker evidence against H0 than does the replication test. For example, when δrep = .28 and norig = nrep = 50, the default test yields an inconclusive Bayes factor of .73. The replication test, however, yields more informative Bayes factors ranging from 1.94 to 4.94 in favor of the alternative hypothesis. This discrepancy between the default JZS Bayes factor test and the Bayes factor replication test is due to the fact that the definition of the alternative hypothesis is more informative in the replication test, featuring a prior that is not centered on zero. Equality-of-Effect-Size Bayes Factor Test Table 3 shows the outcomes for the Bayes factor test for the equality of effect sizes. As expected, the Bayes factor for equality of effect sizes is highest when the effect sizes are equal and sample size is large. As mentioned earlier, compelling evidence for equality of effect sizes can be obtained only for sample sizes that are relatively large. Furthermore, note that the Bayes factor for equality of effect sizes is relatively insensitive to their deviation from zero. For example, within the (norig = nrep = 50) quadrant, consider the entries where δorig = δrep = (.14, .28, .42, .57). The associated Bayes factors from test for equality of effect sizes are B01 = (8.22, 8.05, 8.00, 7.21) – the extent to which the effect sizes deviate from zero does not influence the test, because the test ignores the skeptic’s hypothesis that effect size is zero. In contrast, for those same table entries, the Bayes factor replication test –pitting QUANTIFYING REPLICATION RESULTS JZS Bayes Factor B10 torig δorig norig = 50, nrep = 50 trep 0 1 2 3 4 δrep .00 .14 .28 .42 .57 1 0.14 0.11 0.18 0.73 6.42 100 2 0.28 0.11 0.18 0.73 6.42 100 3 0.42 0.11 0.18 0.73 6.42 100 4 0.57 0.11 0.18 0.73 6.42 100 norig = 100, nrep = 50 trep 0.00 0.71 1.41 2.12 2.83 δrep .0 .1 .2 .3 .4 1 0.1 0.08 0.21 3.32 245.49 87487 2 0.2 0.08 0.21 3.32 245.49 87487 3 0.3 0.08 0.21 3.32 245.49 87487 4 0.4 0.08 0.21 3.32 245.49 87487 14 norig = 50, nrep = 100 0 0.71 1.41 2.12 2.83 .00 .14 .28 .42 .57 0.11 0.14 0.29 0.92 4.22 0.11 0.14 0.29 0.92 4.22 0.11 0.14 0.29 0.92 4.22 0.11 0.14 0.29 0.92 4.22 norig = 100, nrep = 100 0 1 2 3 4 .0 .1 .2 .3 .4 0.08 0.13 0.55 5.64 122.59 0.08 0.13 0.55 5.64 122.59 0.08 0.13 0.55 5.64 122.59 0.08 0.13 0.55 5.64 122.59 Table 2: Results from the independent JZS Bayes factor test B10 applied to replication attempts for different values of torig , trep , norig , and nrep . The rows are defined by the t values in the original study and corresponding effect sizes for original sample sizes of 50 (upper half) and 100 (lower half). The columns are defined by the t values in the replication attempt and corresponding effect sizes for replication sample sizes of 50 (left half) and 100 (right half). Because the test ignores the results from the original study, within each quadrant all rows are identical. the proponent’s Hr against the skeptic’s H0 – yields evidence that greatly increases with the deviation of δ from 0: Br0 = (1.17, 4.94, 46, 770). This qualitative difference between the two tests reflects the fact that they address different substantive questions. Fixed-Effect Meta-Analysis Bayes Factor Test Table 4 shows the outcomes for the fixed-effect meta-analysis Bayes factor. This test pools the data from the original experiment and the replication attempt and determines whether the effect is present or absent overall. As expected, the meta-analytic Bayes factor is relatively high when both effect sizes are large. Whenever one of the effect sizes is low, the other one has to be high to compensate and produce a Bayes factor in favor of the effect. Studies with large sample size have more of an influence on the final outcome than do studies with small sample sizes. In contrast to the new replication test, the meta-analytic Bayes factor treats the data from the original study and the replication attempt as exchangeable. This yields different conclusions, particularly in cases where a high effect size in the original study provides sufficient evidence to compensate for a small effect size in the replication attempt. These situations hold in the lower left corners for each of the quadrants of Table 4. For example, within the (norig = nrep = 50) quadrant, consider the entry where δorig = .57 and δrep = .14. The associated meta-analytic Bayes factor is B10 = 16.28 (i.e., evidence in favor of the overall effect). The same entry yields a replication Bayes factor of Br0 = 0.14 (i.e., evidence against replication: B0r = 1/0.14 = 7.14). Again, this is a qualitative difference that reflects the fact that the tests address different questions and make different assumptions. 15 QUANTIFYING REPLICATION RESULTS Equality-of-Effect-Size Bayes Factor B01 torig δorig norig = 50, nrep = 50 trep 0 1 2 3 δrep .00 .14 .28 .42 1 0.14 6.42 8.22 6.29 2.96 2 0.28 3.12 6.44 8.05 6.27 3 0.42 0.97 3.18 6.44 8.00 4 0.57 0.20 1.01 3.25 6.45 norig = 100, nrep = 50 trep 0.00 0.71 1.41 2.12 δrep .0 .1 .2 .3 1 0.1 7.97 9.34 7.93 4.81 2 0.2 4.92 8.06 9.38 7.84 3 0.3 2.21 4.99 7.92 9.24 4 0.4 0.72 2.22 5.02 7.94 4 .57 1.01 3.14 6.03 7.21 0.00 .00 6.75 2.58 0.54 0.07 2.83 .4 2.12 4.84 7.86 8.79 0 .0 8.89 4.34 1.27 0.24 norig = 50, nrep = 100 0.71 1.41 2.12 .14 .28 .42 9.37 6.79 2.63 6.68 9.26 6.79 2.62 6.66 9.26 0.58 2.66 6.69 norig = 100, nrep = 100 1 2 3 .1 .2 .3 11.44 8.86 4.22 9.07 11.36 8.79 4.32 8.98 11.40 1.33 4.38 9.00 2.83 .57 0.48 2.42 6.44 8.87 4 .4 1.23 4.22 8.72 11.06 Table 3: Results from the Bayes factor for equality of effect sizes B01 applied to different values of torig , trep , norig , and nrep . The rows are defined by the t values in the original study and corresponding effect sizes for original sample sizes of 50 (upper half) and 100 (lower half). The columns are defined by the t values in the replication attempt and corresponding effect sizes for replication sample sizes of 50 (left half) and 100 (right half). Meta-analysis Bayes Factor B10 torig δorig norig = 50, nrep = 50 trep 0 1 2 3 4 δrep .00 .14 .28 .42 .57 1 .14 0.08 0.19 0.68 3.07 16 2 .28 0.19 0.68 3.63 25.97 214 3 .42 0.56 3.07 25.97 295.40 38141 4 .57 1.96 16.28 214.25 3814.01 76447 norig = 100, nrep = 50 trep 0 0.71 1.41 2.12 2.83 δrep .0 .1 .2 .3 .4 1 .1 0.05 0.12 0.23 0.47 1.04 2 .2 0.12 0.49 1.21 3.33 9.87 3 .3 0.44 3.15 10.70 40.55 163.64 4 .4 2.39 31.47 147.06 766.92 4226.75 0 .00 0.05 0.11 0.23 0.52 0.71 .14 0.27 0.83 2.98 11.38 0 .0 0.05 0.14 0.45 1.98 1 .1 0.14 0.50 2.56 17.56 norig = 50, nrep = 100 1.41 2.12 2.83 .28 .42 .57 3.32 100 7175 18.92 1067 148931 124.50 12996 3488338 831.93 154711 77531702 norog = 100, nrep = 100 2 3 4 .2 .3 .4 0.50 2.56 18 2.80 22.92 251 22.92 303.15 5330 251.27 5330.21 149691 Table 4: Results from the meta-analytic Bayes factor the presence or absence of an effect B10 applied to different values of torig , trep , norig , and nrep . The rows are defined by the t values in the original study and corresponding effect sizes for original sample sizes of 50 (upper half) and 100 (lower half). The columns are defined by the t values in the replication attempt and corresponding effect sizes for replication sample sizes of 50 (left half) and 100 (right half) Examples We now apply the above Bayesian t tests to three examples of replication attempts from the literature. These examples cover one-sample and two-sample t tests in which QUANTIFYING REPLICATION RESULTS 16 the outcome is replication failure, replication success, and replication ambivalence. The examples also allow us to visualize the prior and posterior distributions for effect size, as well as the test outcome. The examples were chosen for illustrative purposes and do not reflect our opinion, positive or negative, about the experimental work or the researchers who carried it out. Moreover, our analysis is purely statistical and purposefully ignores the many qualitative aspects that may come into play when assessing the strength of a replication study (see Brandt et al., 2013). For simplicity, we also ignore the possibility that experiments with nonsignificant results may have been suppressed (i.e., publication bias, e.g., Francis, 2013b, in pressa). In each of the examples, it is evident that the graded measure of evidence provided by the Bayesian replication tests is more informative and balanced than the “significant-nonsignificant” dichotomy inherent to the p value assessment of replication success that is currently dominant in psychological research. Example 1: Red, Rank, and Romance in Women Viewing Men In an attempt to unravel the mystery of female sexuality, Elliot et al. (2010) set out to discover what influences women’s attraction to men. Inspired by findings in crustaceans, sticklebacks, and rhesus macaques, Elliot et al. (2010) decided to test the hypothesis that “viewing red leads women to perceive men as more attractive and more sexually desirable” (p. 400). In a series of experiments, female undergraduate students were shown a picture of a moderately attractive man; subsequently, the students had to indicate their perception of the man’s attractiveness. The variable of interest was either the man’s shirt color or the picture background color (for a critique see Francis, 2013a). The first experiment of Elliot et al. (2010) produced a significant effect of color on perceived attractiveness (t(20) = 2.18, p < .05, δ = 0.95): the female students rated the target man as more attractive when the picture was presented on a red background (M = 6.79, SD = 1.00) than when it was presented on a white background (M = 5.67, SD = 1.34). The second experiment was designed to replicate the first and to assess whether the color effect generalised to male students. The results showed the predicted effect of color on perceived attractiveness for female students (t(53) = 3.06, p < .01, δ = 1.11) but not for male students (t(53) = 0.25, p > .80, δ = .03). In the third experiment, featuring female participants only, the neutral background color was changed from white to grey. The results again confirmed the presence of the predicted color effect (t(32) = 2.44, p < .05, δ = 0.86): the students rated the target man as more attractive when the picture was presented on a red background (M = 6.69, SD = 1.22) than when it was presented on a grey background (M = 5.27, SD = 2.04). We now re-analyze these results with our Bayesian replication tests, assuming that Experiment 1 from Elliot et al. (2010) is the original study and the others are the replication attempts. The results are summarized in Table 5.7 Before discussing the extant Bayes factor tests we first discuss and visualize the results from our new Bayes factor test for replication. The left panel of Figure 2 shows the results for the data from the female students in Elliot et al.’s Experiment 2. The dotted line indicates the prior distribution for the replication test, that is, the proponent’s posterior distribution for effect size after observing the data from the original experiment, p(δ | Yorig ). The solid 7 R code to reproduce the analyses for all of the examples is available from the first author’s webpage. 17 QUANTIFYING REPLICATION RESULTS Rep Br0 Red and Romance Original δ = 0.95 Female δ = 1.33 Male δ = 0.11 Gray δ = 1.07 Professor priming Original δ = 0.91 Replication 1 δ = 0.14 Replication 2 δ = -0.08 JZS B10 (one-sided) Equality B01 Meta B10 3.09 (489*) 39.73 0.13 9.76 1.79 10.51 0.21 2.75 0.16 0.05 8.92 0.24 0.22 (20.96) (0.25) (5.42) 3.29 1.07 3.10 0.21 (0.32) (0.18) Negative priming Original δ = 0.74 10.45 Replication 1 δ = 0.46 2.37 0.97 (1.89) Replication 2 δ = -0.54 0.01 1.67 (0.04) * Result for the meta-analysis without the Male study. 0.99 0.39 0.11 2.95 0.00 Table 5: Results for four different Bayes factor tests applied to three example studies. Bayes factors higher than 1 favor the hypothesis that an effect is present. Note: “Rep B10 ” is the new Bayes factor test for replication; “JZS B10 ” is the two-sided independent default Bayes factor test, with the result of the one-sided test between brackets; “Equality B01 ” is the equality-of-effect-size Bayes factor test; and “Meta B10 ” is the fixed-effect meta-analysis Bayes factor test. line indicates the posterior distribution for effect size after observing the additional data from the replication experiment, p(δ | Yrep , Yorig ). This posterior distribution assigns more mass to values of δ higher than zero than did the prior distribution, a finding that suggests replication success. The extent of this success is quantified by the computation of Br0 , which yields 39.73: the data from the female students in Experiment 2 are about 40 times more likely under the proponent’s replication hypothesis Hr than under the skeptic’s null hypothesis H0 . The outcome of the new Bayesian replication t test is visualized by the ordinates of the prior and posterior distributions at δ = 0, indicated in the left panel of Figure 2 by the two filled circles. Intuitively, the height of the prior distribution at δ = 0 reflects the believability of H0 before seeing the data from the replication attempt, and the height of the posterior distribution at δ = 0 reflects the believability of H0 after seeing those data. In this case, observing the replication data decreases the believability of H0 . It so happens that the ratio of the ordinates exactly equals the Bayes factor Br0 (i.e., the Savage-Dickey density ratio test, see e.g., Dickey & Lientz, 1970; Wagenmakers, Lodewyckx, Kuriyal, & Grasman, 2010), providing a visual confirmation of the test outcome. The middle panel of Figure 2 shows the results for the data from the male students in Elliot et al.’s Experiment 2. The posterior distribution assigns more mass to δ = 0 than does the prior distribution, indicating evidence in favor of the null hypothesis H0 over the replication hypothesis Hr . The Bayesian replication test yields Br0 = 0.13, indicating that QUANTIFYING REPLICATION RESULTS 18 the replication data are 1/.13 = 7.69 times more likely under H0 than under Hr . Finally, the right panel of Figure 2 shows the results for the data from Elliot et al.’s Experiment 3. After observing the data from this replication attempt the posterior distribution assigns more mass to values of δ higher than zero than did the prior distribution; hence the Bayesian replication test indicates support for the replication hypothesis Hr over the null hypothesis H0 . The Bayes factor equals 9.76, indicating that the data are almost 10 times more likely to occur under Hr than under H0 . Because the sample size in this replication attempt is smaller than in the first replication attempt, the posterior is less peaked and the test outcome less extreme. Figure 2. Results from the Bayes factor replication test applied to Experiment 2 and 3 from Elliot et al. (2010). The original experiment had shown that female students judge men in red to be more attractive. The left and middle panel show the results from Elliot et al.’s Experiment 2 (for female and male students, respectively), and the right panel shows the results from Elliot et al.’s Experiment 3 that featured a different control condition. In each panel, the dotted lines represent the posterior from the original experiment, which is used as prior for effect size in the replication tests. The solid lines represent the posterior distributions after the data from the replication attempt are taken into account. The grey dots indicate the ordinates of this prior and posterior at the skeptic’s null hypothesis that the effect size is zero. The ratio of these two ordinates gives the result of the replication test. In contrast to the analysis above, the three extant Bayes factor tests produce results that are more unequivocal. The default JZS Bayes factor equals 1.79 for the original experiment, indicating that the data are uninformative, as they are almost as likely to have occurred under the null hypothesis as under the alternative hypothesis. The study with female students yields strong support in favor of the presence of an effect (B10 = 10.51), whereas the study with male students yields moderate evidence in favor of the absence of an effect (B01 = 1/B10 = 1/0.21 = 4.76). The study with the gray background, however, yields only anecdotal support for the presence of an effect (B10 = 2.75; a one-sided test, however, almost doubles the support, B10 = 5.42). The equality-of-effect-size Bayes factor test does not yield strong support for or against the hypothesis that the effect sizes are equal, with Bayes factors that never exceed 3.29. The fixed-effect meta-analysis Bayes factor test that pools the data from all four studies yields B10 = 3.09 in favor of the presence of an overall effect, including the hypothesis that the effect generalises to men. When the experiment with male students is omitted the pooled QUANTIFYING REPLICATION RESULTS 19 Bayes factor is B10 = 489, indicating extreme evidence in favor of an effect. Example 2: Dissimilarity Contrast in the Professor Priming Effect Seminal work by Dijksterhuis and others (e.g., Dijksterhuis & van Knippenberg, 1998; Dijksterhuis et al., 1998) has suggested that priming people with intelligence-related concepts (e.g., “professor”) can make them behave more intelligently (e.g., answer more trivial questions correctly). The extent to which this effect manifests itself is thought to depend on whether the prime results in assimilation or contrast (e.g., Mussweiler, 2003). Specifically, the presentation of general categories such as “professor” may lead to assimilation (i.e., activation of the concept of intelligence), whereas the presentation of a specific exemplar such as “Einstein” may lead to contrast (i.e., activation of the concept of stupidity). However, LeBoeuf and Estes (2004) argued that the balance between assimilation and contrast is determined not primarily by whether the prime is a category or an exemplar, but rather by whether the prime is perceived as relevant in the social comparison process. To test their hypothesis, LeBoeuf and Estes (2004) designed an experiment in which different groups of participants were first presented with either the category prime “professor” or the exemplar prime “Einstein”. To manipulate prime relevance participants were then asked to list similarities and differences between themselves and the presented prime. Subsequently, a test phase featured a series of multiple-choice general knowledge questions. The results showed that performance was better in the difference-listing condition than in the similarity-listing condition. LeBoeuf and Estes (2004) interpreted these findings as follows: “As hypothesized, when participants were encouraged to perceive themselves as unlike a prime, behavior assimilated to prime activated traits, presumably because the prime was rejected as a comparison standard. When participants contemplated how they were similar to a prime, that prime was seemingly adopted as a relevant standard for selfcomparison. With the current primes, such a comparison led to negative selfevaluations of intelligence and to lower test performance. Counterintuitively, participants who considered similarities between themselves and an intelligent prime exhibited worse performance than did participants who considered differences between themselves and an intelligent prime.” (pp. 616-617, italics in original). Among the various conditions of Experiment 1 in LeBoeuf and Estes (2004), the best performance was attained in the “differences to Einstein” condition (56.2% correct) whereas the worst performance was observed in the “similarities to professors” condition (45.2%), a performance gap that was highly significant (t(42) = 3.00, p = .005, δ = .45 ). These two cells in the design were the target of two recent replication attempts by Shanks et al. (2013). Specifically, Experiment 5 from Shanks et al. (2013) was designed to be “as close to LeBoeuf and Estes’ study as possible” and yielded a nonsignificant effect that was slightly in the opposite direction (t(47) = −.25, p = .60, δ = −.03, one-sided t test). Experiment 6 from Shanks et al. (2013) tried to maximize the possibility of finding the effect by informing participants beforehand about the hypothesized effects of the primes; however, the results again yielded a nonsignificant effect that was slightly in the opposite direction (t(30) = −1.25, p = .89, δ = −.22, one-sided t test). We now re-analyze these results, assuming that Experiment 1 from LeBoeuf and Estes (2004) is the original study and Experiment 5 and 6 from Shanks et al. (2013) are the replication attempts. The resulting Bayes factors are presented in Table 5. QUANTIFYING REPLICATION RESULTS 20 Before discussing the extant Bayes factor tests we first discuss and visualize the results from our new Bayes factor test for replication. In both panels of Figure 3, the dotted line indicates the prior distribution for the replication test, that is, the proponent’s posterior distribution for effect size after observing the data from the original experiment, p(δ | Yorig ). The left panel of Figure 3 shows the results for the data from Shanks et al.’s Experiment 5. The solid line indicates the posterior distribution for effect size after observing the additional data from the replication experiment, p(δ | Yrep , Yorig ). The posterior distribution assigns more mass to δ = 0 than does the prior distribution, indicating evidence in favor of the null hypothesis H0 over the replication hypothesis Hr . The Bayesian replication test yields Br0 = 0.23, indicating that the replication data are 1/.23 = 4.35 times more likely under H0 than under Hr . This constitutes some evidence against Hr , although it is relatively weak and only just exceeds Jeffreys’ threshold for evidence that is “not worth more than a bare mention”. The right panel of Figure 3 shows the results for the data from Shanks et al.’s Experiment 6. The solid line again indicates the posterior distribution after observing the data from the replication experiment. As in the left panel, the posterior distribution assigns more mass to δ = 0 than does the prior distribution, and the Bayesian replication test yields Br0 = 0.05, indicating that the replication data are 1/.05 = 20 times more likely under H0 than under Hr . This constitutes strong evidence against Hr . Figure 3. Results from the Bayes factor replication test applied to Experiment 5 (left panel) and Experiment 6 (right panel) from Shanks et al. (2013). In each panel, the dotted lines represent the posterior from the original experiment by LeBoeuf and Estes (2004), which is used as prior for effect size in the replication tests. The solid lines represent the posterior distributions after the data from the replication attempt are taken into account. The grey dots indicate the ordinates of this prior and posterior at the skeptic’s null hypothesis that the effect size is zero. The ratio of these two ordinates gives the result of the replication test. The results from the three extant Bayes factor tests are as follows. The default JZS Bayes factor equals 8.92 for the original experiment, indicating moderate to strong support for the alternative hypothesis. The two replication attempts, however, show the opposite pattern. In both studies, the data are about five times more likely under the null hypothesis QUANTIFYING REPLICATION RESULTS 21 than under the alternative hypothesis. The equality-of-effect-size Bayes factor test does not yield strong support for or against the hypothesis that the effect sizes are equal: for the first replication attempt the outcome is almost completely uninformative, and for the second replication attempt the data are only 2.75 (1/.39) times more likely under the hypothesis that the effect sizes are unequal rather than equal. The fixed-effect meta-analysis Bayes factor test that pools the data from all three studies yields B10 = 0.21 in favor of the presence of an overall effect, which indicates that the combined data are about five times (i.e., 1/.21) more likely under the null hypothesis of no effect. Example 3: Negative Priming Negative priming refers to the decrease in performance (e.g., longer response times or RTs, more errors) for a stimulus that was previously presented in a context in which it had to be ignored. For instance, assume that on each trial of an experiment, participants are confronted with two words: one printed in red, the other in green. Participants are told to respond only to the red target word (e.g., by indicating whether or not it represents an animate entity, horse: yes, furnace: no) and ignore the green distractor word. Negative priming is said to have occurred when performance on the target word suffers when, on the previous trial, this same word was presented as a green distractor. The theoretical relevance of negative priming is that it is evidence for inhibitory processing – commonly, negative priming is attributed to an attentional mechanism that actively suppresses or inhibits irrelevant stimuli; when these inhibited stimuli later become relevant and need to be re-activated, this process takes time and a performance decrement is observed. However, the standard suppression account of negative priming was called into question by Milliken et al. (1998), who showed that negative priming can also be observed in situations where the first presentation of the repeated item does not specifically call for it to be ignored. Specifically, we focus here on Experiment 2A from Milliken et al. (1998), where participants had to name a target word printed in red, while ignoring a distractor word printed in green. Prior to the target stimulus display, a prime word was presented for 33 ms, printed in white. No action was required related to the prime word. In the unrepeated condition, the prime was not related to the words from the target display; in the repeated condition, the prime was identical to the target word that was presented 500 ms later. Despite the fact that the prime was so briefly presented that most participants were unaware of its presence, and despite the fact that no active suppression of the prime was called for, Milliken et al. (1998) nonetheless found evidence for negative priming: RTs were 8 ms slower in the repeated than in the unrepeated condition, t(19) = 3.29, p < .004, δ = .74. The experiment by Milliken et al. (1998) was the target for two nearly exact replication attempts. In Experiment 1A from Neill and Kahan (1999), the negative priming effect was again observed: RTs in the repeated condition were 13 ms slower than those in the unrepeated condition, t(29) = 2.06, p = .048, δ = .38. However, the results from Experiment 1B showed the opposite result, with RTs in the repeated condition being 7 ms faster than those in the unrepeated condition, t(43) = −2.40, p = .021, δ = .37. We now re-analyze these results with our Bayesian replication t test, assuming that Experiment 2A from Milliken et al. (1998) is the original study and Experiments 1A and QUANTIFYING REPLICATION RESULTS 22 1B from Neill and Kahan (1999) are the replication attempts. The results are presented in Table 5. Before discussing the extant Bayes factor tests we first discuss and visualize the results from our new Bayes factor test for replication. Similar to the previous examples, in both panels of Figure 4 the dotted line indicates the prior distribution for the replication test, that is, the proponent’s posterior distribution for effect size after observing the data from the original experiment, p(δ | Yorig ). The left panel of Figure 4 shows the results for the data from Neill and Kahan’s Experiment 1A. The solid line indicates the posterior distribution for effect size after observing the additional data from the replication experiment, p(δ | Yrep , Yorig ). Although barely discernable from the plot, the posterior distribution assigns less mass to δ = 0 than does the prior distribution, indicating evidence in favor of the replication hypothesis Hr over the null hypothesis H0 . The Bayesian replication test yields Br0 = 2.37, indicating that the replication data are 2.37 times more likely under Hr than under H0 . This constitutes evidence in favor of the replication hypothesis Hr , albeit weak and, according to Jeffreys, “not worth more than a bare mention”. The right panel of Figure 4 shows the results for the data from Neill and Kahan’s Experiment 1B. The solid line again indicates the posterior distribution after observing the data from the replication experiment. Now the posterior distribution assigns much more mass to δ = 0 than does the prior distribution, and the Bayesian replication test yields Br0 = 0.01, indicating that the replication data are 1/.01 = 100 times more likely under H0 than under Hr . This constitutes compelling evidence against Hr . Figure 4. Results from the Bayes factor replication test applied to the replication studies of Neill and Kahan (1999) 1A (left panel) and 1B (right panel). In each panel, the dotted lines represent the posterior from the original experiment by Milliken et al. (1998), which is used as prior for effect size in the replication tests. The solid lines represent the posterior distributions after the data from the replication attempt are taken into account. The grey dots indicate the ordinates of this prior and posterior at the skeptic’s null hypothesis that the effect size is zero. The ratio of these two ordinates gives the result of the replication test. QUANTIFYING REPLICATION RESULTS 23 The results from the three extant Bayes factor tests are as follows. The default JZS Bayes factor equals 10.45 for the original experiment, indicating strong support for the alternative hypothesis. The results for the two replication attempts, however, are unequivocal. In both studies, the data are about equally likely under the null hypothesis as under the alternative hypothesis. For the second replication attempt, this outcome is artificial, brought about by the fact that the analyses are two-sided instead of one-sided. A one-sided default Bayes factor test for the second replication attempt provides strong evidence in favor of the null hypothesis versus the one-sided alternative hypothesis that postulates a decrease in performance in the repeated condition, B01 = 26.87 (Morey & Wagenmakers, 2014; Wagenmakers et al., 2010). For the first replication attempt, the equality-of-effect-size Bayes factor test provides anecdotal support for the equality of effect sizes; however, for the second replication attempt the test strongly supports the hypothesis of unequal effect sizes – consistent with the fact that the effect is in the opposite direction. Finally, the fixed-effect meta-analysis Bayes factor test that pools the data from all three studies yields B10 = 0.11 in favor of the presence of an overall effect, which indicates that the combined data are about ten times (1/.11) more likely under the null hypothesis of no effect. Discussion Inspired by recent work from Simonsohn (2013) and Dienes (2011), we proposed a Bayesian replication test to quantify the extent to which a replication attempt has succeeded or failed. The test assesses the relative adequacy of two hypotheses: the skeptic’s null hypothesis and the proponent’s replication hypothesis. The replication hypothesis is formalized as the posterior distribution on effect size that is obtained from the original experiment. Our Bayesian replication test has several advantages over existing methods. First, our test summarizes the proponent’s belief by an entire distribution of effect sizes, not just a point estimate; this way, the sample size of the original experiment is fully taken into account. Second, our test does not require the user to specify prior distributions, immunizing it against the popular objection that Bayes factor hypothesis tests are overly sensitive to the choice of prior distributions (Kruschke, 2010; Liu & Aitkin, 2008; but see Vanpaemel, 2010). Third, by taking into account the size of the original effect, the test is optimally sensitive to detect replication effects of the same size as observed in the original study. In addition, there are the general advantages of Bayes factors over classical p value procedures (e.g., Wagenmakers, Lee, Lodewyckx, & Iverson, 2008); for instance, Bayes factors do not assign special status to the null hypothesis, Bayes factors promote quantitative thinking by providing a continuous measure of support, Bayes factors allow evidence to be monitored as the data come in (Rouder, in press; Wagenmakers et al., 2012), and Bayes factors allow researchers to quantify evidence in favor of the null hypothesis (e.g., Gallistel, 2009). Despite these advantages, it is clear that no single test can address the multitude of questions that pertain to the analysis of data from a replication attempt. Instead, a comprehensive and coherent replication report may consist of several complementary Bayesian analysis procedures in addition to the replication test proposed here. Specifically, the default JZS Bayes factor hypothesis test assesses the evidence for the presence of an effect in the replication study, independent of the outcome of the original experiment. This test QUANTIFYING REPLICATION RESULTS 24 is useful to detect effects that are present but much smaller in magnitude than those from the original experiment. The equality-of-effect-size Bayes factor hypothesis test assesses the evidence for and against the hypothesis that the effects in the original experiment and the replication attempt are of the same size. Finally, the fixed-effect meta-analysis Bayes factor hypothesis test pools the data and assesses the evidence for and against the hypothesis that there exists an effect overall. Together, the suite of Bayesian tests address important but slightly different questions about the result of a replication attempt, namely: “Is the effect similar to what was found before, or is it absent?” (i.e., the replication test), “Is the effect present or absent?” (i.e., the JZS test), “Is the effect equally large as the effect that was found before, or is it different” (i.e., the equality-of-effect-size test), and “do the pooled studies provide evidence for the presence or absence of an effect?” (i.e., the fixed-effect meta-analysis test). As outlined earlier, each test has its own assumptions and, consequently, its own drawbacks. For instance, a skeptic who has ground to doubt the findings from an original experiment will be reluctant to use the equality-of-effect-size test, because this test ignores the possibility that the effect is absent. The skeptic’s reluctance will be even greater towards the fixed-effect meta-analysis test, because this test assumes that which the skeptic wishes to dispute (i.e., that the data from the original experiment and the replication experiment are exchangeable). On the other hand, when the original experiment did not find an effect, there is no “proponent” and instead of our replication test it may be more appropriate to apply the equality-of-effect-size test. The importance of model assumptions is further underscored by comparing the replication test to the default JZS test. For instance, when the original experiment finds an effect size of about .9, and the replication finds a much smaller effect size of about .3, our replication test may indicate that the replication attempt has failed. Application of the default Bayes factor hypothesis test, however, may indicate that both the original and the replication experiment provide strong evidence against H0 . This is not paradoxical – it just indicates that the different hypothesis tests address different questions. As mentioned earlier, our replication test addresses the question “Is the effect from the replication attempt comparable to what was found before, or is it zero?” whereas the default test addresses the question “Given that we know relatively little about the expected effect size beforehand, is the effect from the replication attempt present or absent?”. Similarly, suppose that the original experiment finds an effect size of about .3, and the replication finds a much larger effect size of .9. It is likely that the replication test will indicate strong support for Hr . Again, this is not paradoxical – it is true that Hr provides a poor account of the data from the replication attempt, but H0 will do even worse. Extensions and Future Directions The suite of Bayesian methods outlined here is based on a framework of hypothesis testing, where different statistical accounts of the data are contrasted and formally evaluated. However, in addition to hypothesis testing one may also use parameter estimation as a method of inference (e.g., Cumming, in press, but see Morey, Rouder, Verhagen, & Wagenmakers, in press). As argued by Berger (2006), a complete assessment includes both results from hypothesis testing and from parameter estimation. Within the Bayesian framework, parameter estimation centers on the posterior distribution, the distribution of uncertainty QUANTIFYING REPLICATION RESULTS 25 or degree of belief about the parameter under consideration. For replication research, it may be informative to show the posterior distribution for effect size, both for the original experiment and the replication attempt. The question addressed by such parameter estimation procedures is “If the effect is present, how big is it?”. A related method is to show the posterior distribution of the difference in effect size between the original experiment and the replication attempt (e.g., Bayarri & Mayoral, 2002a, 2002b). The replication t test proposed here is a first step in a larger research agenda; we are currently working to extend the test to other popular hypothesis tests such as analysis of variance (ANOVA) and the test for the presence of a correlation. Our test may be particularly appropriate when skeptics try to ascertain whether or not a replication attempt is consistent with an earlier finding. Our test may also be useful for the analysis of replication attempts from the Open Science Framework project on reproducibility. Proponents could also use our replication test – for instance, they could use the test to quantify the extent to which a follow-up study successfully replicated an earlier study. To prevent the results from being contaminated by hindsight bias and publication bias, we recommend that replication studies –and perhaps all studies– are preregistered so that it is possible to distinguish clearly between work that is hypothesis-generating versus hypothesis-testing (De Groot, 1956/in press; Wagenmakers et al., 2012). The replication test can also be extended by quantifying the evidential impact of multiple replication attempts. For instance, a second replication attempt need not use the same prior as the first replication attempt. Instead, the prior for the assessment of the second replication attempt can be the rational compromise between the prior for the first replication experiment and the data from the first replication attempt. Our replication test is conceptually similar to the partial Bayes factor (Berger & Pericchi, 1996; O’Hagan & Forster, 2004; Lempers, 1971) in which a small part of the data is used as a training set in order to update a vague prior distribution, and the remaining part of the data is used for Bayes factor model comparison. In our method, the training sample consists of the data from the original experiment, and the remaining part corresponds to data from the replication attempt. As the emphasis towards transparency and replications in psychology continues to increase, we hope for a concomitant raise in awareness concerning the proper statistical techniques for drawing conclusions from data. Fortunately, the findings in our field are often sufficiently clear as to pass what Berkson called the interocular traumatic test – when the data are so compelling that conclusion hits you straight between the eyes. However, “...the enthusiast’s interocular trauma may be the skeptic’s random error. A little arithmetic to verify the extent of the trauma can yield great peace of mind for little cost.” (Edwards et al., 1963, p. 217). This is precisely the goal of our replication test: to “verify the extent of the trauma” from a replication attempt, and to do so in a way that is intuitive to understand, straightforward to apply, and easy to interpret. QUANTIFYING REPLICATION RESULTS 26 References Asendorpf, J. B., Conner, M., De Fruyt, F., De Houwer, J., Denissen, J. J. A., Fiedler, K., Fiedler, S., Funder, D. C., Kliegl, R., Nosek, B. A., Perugini, M., Roberts, B. W., Schmitt, M., van Aken, M. A. G., Weber, H., & Wicherts, J. M. (2013). Recommendations for increasing replicability in psychology. European Journal of Personality, 27, 108–119. Bayarri, M. J., Berger, J. O., Forte, A., & Garcı́a-Donato, G. (2012). Criteria for Bayesian model choice with application to variable selection. The Annals of Statistics, 40, 1550–1577. Bayarri, M. J., & Garcı́a-Donato, G. (2007). Extending conventional priors for testing general hypotheses in linear models. Biometrika, 94, 135–152. Bayarri, M. J., & Mayoral, A. M. (2002a). Bayesian analysis and design for comparison of effect– sizes. Journal of Statistical Planning and Inference, 103, 225–243. Bayarri, M. J., & Mayoral, A. M. (2002b). Bayesian design of “successful” replications. The American Statistician, 56, 207–214. Bem, D. J., Utts, J., & Johnson, W. O. (2011). Must psychologists change the way they analyze their data? Journal of Personality and Social Psychology, 101, 716–719. Berger, J. O. (2006). Bayes factors. In S. Kotz, N. Balakrishnan, C. Read, B. Vidakovic, & N. L. Johnson (Eds.), Encyclopedia of statistical sciences, vol. 1 (2nd ed.) (pp. 378–386). Hoboken, NJ: Wiley. Berger, J. O., & Pericchi, L. R. (1996). The intrinsic Bayes factor for model selection and prediction. Journal of the American Statistical Association, 91, 109–122. Bones, A. K. (2012). We knew the future all along: Scientific hypothesizing is much more accurate than other forms of precognition–a satire in one part. Perspectives on Psychological Science, 7, 307–309. Brandt, M. J., IJzerman, H., Dijksterhuis, A., Farach, F. J., Geller, J., Giner-Sorolla, R., Grange, J. A., Perugini, M., Spies, J. R., & van ’t Veer, A. (2013). The replication recipe: What makes for a convincing replication? Manuscript submitted for publication. Cohen, J. (1988). Statistical power analysis for the behavioral sciences. Hillsdale, NJ: Erlbaum. Cohen, J. (1990). Things I have learned (so far). American Psychologist, 45, 1304–1312. Cumming, G. (in press). The new statistics: Why and how. Psychological Science. Cumming, G., & Finch, S. (2001). A primer on the understanding, use, and calculation of confidence intervals that are based on central and noncentral distributions. Educational and Psychological Measurement, 61, 532-574. De Groot, A. D. (1956/in press). The meaning of “significance” for different types of research. Translated and annotated by Eric-Jan Wagenmakers, Denny Borsboom, Josine Verhagen, Rogier Kievit, Marjan Bakker, Angelique Cramer, Dora Matzke, Don Mellenbergh, and Han L. J. van der Maas. Acta Psychologica. Dickey, J. M., & Lientz, B. P. (1970). The weighted likelihood ratio, sharp hypotheses about chances, the order of a Markov chain. The Annals of Mathematical Statistics, 41, 214–226. Dienes, Z. (2011). Bayesian versus orthodox statistics: Which side are you on? Perspectives on Psychological Science, 6, 274–290. Dijksterhuis, A., Spears, R., Postmes, T., Stapel, D., Koomen, W., van Knippenberg, A., & Scheepers, D. (1998). Seeing one thing and doing another: Contrast effects in automatic behavior. Journal of Personality and Social Psychology, 75, 862–871. QUANTIFYING REPLICATION RESULTS 27 Dijksterhuis, A., & van Knippenberg, A. (1998). The relation between perception and behavior, or how to win a game of Trivial Pursuit. Journal of Personality and Social Psychology, 74, 865–877. Edwards, W., Lindman, H., & Savage, L. J. (1963). Bayesian statistical inference for psychological research. Psychological Review, 70, 193–242. Efron, B., & Tibshirani, R. J. (1993). An introduction to the bootstrap. New York: Chapman & Hall. Elliot, A. J., Kayser, D. N., Greitemeyer, T., Lichtenfeld, S., Gramzow, R. H., Maier, M. A., & Liu, H. (2010). Red, rank, and romance in women viewing men. Journal of Experimental Psychology: General, 139, 399–417. Francis, G. (2013a). Publication bias in “Red, rank, and romance in women viewing men,” by Elliot et al. (2010). Journal of Experimental Psychology: General, 142, 292–296. Francis, G. (2013b). Replication, statistical consistency, and publication bias. Journal of Mathematical Psychology, 57, 153–169. Francis, G. (in pressa). The frequency of excess success for articles in Psychological Science. Psychonomic Bulletin & Review. Francis, G. (in pressb). Replication, statistical consistency, and publication bias. Journal of Mathematical Psychology. Gallistel, C. R. (2009). The importance of proving the null. Psychological Review, 116, 439–453. Gelman, A., & Stern, H. (2006). The difference between “significant” and “not significant” is not itself statistically significant. The American Statistician, 60, 328–331. Jeffreys, H. (1961). Theory of probability (3 ed.). Oxford, UK: Oxford University Press. John, L. K., Loewenstein, G., & Prelec, D. (2012). Measuring the prevalence of questionable research practices with incentives for truth–telling. Psychological Science, 23, 524–532. Kass, R. E., & Raftery, A. E. (1995). Bayes factors. Journal of the American Statistical Association, 90, 773–795. Kruschke, J. K. (2010). Doing Bayesian data analysis: A tutorial introduction with R and BUGS. Burlington, MA: Academic Press. LeBel, E. P., & Peters, K. R. (2011). Fearing the future of empirical psychology: Bem’s (2011) evidence of psi as a case study of deficiencies in modal research practice. Review of General Psychology, 15, 371–379. LeBoeuf, R. A., & Estes, Z. (2004). “Fortunately, I’m no Einstein”: Comparison relevance as a determinant of behavioral assimilation and contrast. Social Cognition, 22, 607–636. Lecoutre, B. (1999). Two useful distributions for Bayesian predictive procedures under normal models. Journal of Statistical Planning and Inference, 79, 93–105. Lecoutre, B. (2007). Another look at confidence intervals for the noncentral T distribution. Journal of Modern Applied Statistical Methods, 6, 107–116. Lee, M. D., & Wagenmakers, E.-J. (2013). Bayesian modeling for cognitive science: A practical course. Cambridge University Press. Lempers, F. B. (1971). Posterior probabilities of alternative linear models. Rotterdam: University Press. QUANTIFYING REPLICATION RESULTS 28 Liang, F., Paulo, R., Molina, G., Clyde, M. A., & Berger, J. O. (2008). Mixtures of g priors for Bayesian variable selection. Journal of the American Statistical Association, 103, 410–423. Lindley, D. V. (1965). Introduction to probability & statistics from a Bayesian viewpoint. Part 2. Inference. Cambridge: Cambridge University Press. Liu, C. C., & Aitkin, M. (2008). Bayes factors: Prior sensitivity and model generalizability. Journal of Mathematical Psychology, 52, 362–375. Milliken, B., Joordens, S., Merikle, P. M., & Seiffert, A. E. (1998). Selective attention: A reevaluation of the implications of negative priming. Psychological Review, 105, 203–229. Morey, R. D., & Rouder, J. N. (2011). Bayes factor approaches for testing interval null hypotheses. Psychological Methods, 16, 406–419. Morey, R. D., Rouder, J. N., Verhagen, J., & Wagenmakers, E.-J. (in press). Why hypothesis tests are essential for psychological science: A comment on Cumming. Psychological Science. Morey, R. D., & Wagenmakers, E.-J. (2014). Simple relation between Bayesian order-restricted and point-null hypothesis tests. Manuscript submitted for publication. Mussweiler, T. (2003). Comparison processes in social judgment: Mechanisms and consequences. Psychological Review, 110, 472–489. Neill, W. T., & Kahan, T. A. (1999). Response conflict reverses priming: A replication. Psychonomic Bulletin & Review, 6, 304–308. Neuroskeptic. (2012). The nine circles of scientific hell. Perspectives on Psychological Science, 7, 643–644. Nieuwenhuis, S., Forstmann, B. U., & Wagenmakers, E.-J. (2011). Erroneous analyses of interactions in neuroscience: A problem of significance. Nature Neuroscience, 14, 1105–1107. Nosek, B. A., & Bar–Anan, Y. (2012). Scientific utopia: I. Opening scientific communication. Psychological Inquiry, 23, 217–243. Nosek, B. A., & Lakens, D. (2013). Call for proposals: Special issue of social psychology on “replications of important results in social psychology”. Social Psychology, 44, 59–60. Nosek, B. A., Spies, J. R., & Motyl, M. (2012). Scientific utopia: II. Restructuring incentives and practices to promote truth over publishability. Perspectives on Psychological Science, 7, 615–631. O’Hagan, A., & Forster, J. (2004). Kendall’s advanced theory of statistics vol. 2B: Bayesian inference (2nd ed.). London: Arnold. Pashler, H., & Wagenmakers, E.-J. (2012). Editors’ introduction to the special section on replicability in psychological science: A crisis of confidence? Perspectives on Psychological Science, 7, 528– 530. Rosenthal, R., & Rubin, D. B. (1990). Replication in behavioral research. Journal of Social Behavior and Personality, 5, 1–30. Rouanet, H. (1996). Bayesian methods for assessing importance of effects. Psychological Bulletin, 119, 149-158. Rouder, J. N. (in press). Optional stopping: No problem for Bayesians. Psychonomic Bulletin & Review. Rouder, J. N., & Morey, R. D. (2011). A Bayes–factor meta analysis of Bem’s ESP claim. Psychonomic Bulletin & Review, 18, 682–689. QUANTIFYING REPLICATION RESULTS 29 Rouder, J. N., & Morey, R. D. (2012). Default Bayes factors for model selection in regression. Multivariate Behavioral Research, 47, 877–903. Rouder, J. N., Morey, R. D., Speckman, P. L., & Province, J. M. (2012). Default Bayes factors for ANOVA designs. Journal of Mathematical Psychology, 56, 356–374. Rouder, J. N., Speckman, P. L., Sun, D., Morey, R. D., & Iverson, G. (2009). Bayesian t tests for accepting and rejecting the null hypothesis. Psychonomic Bulletin & Review, 16, 225–237. Royall, R. M. (1997). Statistical evidence: A likelihood paradigm. London: Chapman & Hall. Sellke, T., Bayarri, M. J., & Berger, J. O. (2001). Calibration of p values for testing precise null hypotheses. The American Statistician, 55, 62–71. Shanks, D. R., Newell, B. R., Lee, E. H., Balakrishnan, D., Ekelund, L., Cenac, Z., Kavvadia, F., & Moore, C. (2013). Priming intelligent behavior: An elusive phenomenon. PLoS ONE, 8, e56515. Simmons, J. P., Nelson, L. D., & Simonsohn, U. (2011). False–positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychological Science, 22, 1359–1366. Simonsohn, U. (2013). Evaluating replication results. Manuscript submitted for publication. Ullrich, J., Krueger, J., Brod, A., & Groschupf, F. (in press). More is not less: Greater information quantity does not diminish liking. Journal of Personality and Social Psychology. Vanpaemel, W. (2010). Prior sensitivity in theory testing: An apologia for the Bayes factor. Journal of Mathematical Psychology, 54, 491–498. Wagenmakers, E.-J., Lee, M. D., Lodewyckx, T., & Iverson, G. (2008). Bayesian versus frequentist inference. In H. Hoijtink, I. Klugkist, & P. A. Boelen (Eds.), Bayesian evaluation of informative hypotheses (pp. 181–207). New York: Springer Verlag. Wagenmakers, E.-J., Lodewyckx, T., Kuriyal, H., & Grasman, R. (2010). Bayesian hypothesis testing for psychologists: A tutorial on the Savage–Dickey method. Cognitive Psychology, 60, 158–189. Wagenmakers, E.-J., van der Maas, H. J. L., & Grasman, R. P. P. P. (2007). An EZ–diffusion model for response time and accuracy. Psychonomic Bulletin & Review, 14, 3–22. Wagenmakers, E.-J., Wetzels, R., Borsboom, D., & van der Maas, H. L. J. (2011). Why psychologists must change the way they analyze their data: The case of psi. Journal of Personality and Social Psychology, 100, 426–432. Wagenmakers, E.-J., Wetzels, R., Borsboom, D., van der Maas, H. L. J., & Kievit, R. A. (2012). An agenda for purely confirmatory research. Perspectives on Psychological Science, 7, 627–633. Wetzels, R., Grasman, R. P. P. P., & Wagenmakers, E.-J. (2012). A default Bayesian hypothesis test for ANOVA designs. The American Statistician, 66, 104–111. Wetzels, R., Matzke, D., Lee, M. D., Rouder, J. N., Iverson, G. J., & Wagenmakers, E.-J. (2011). Statistical evidence in experimental psychology: An empirical comparison using 855 t tests. Perspectives on Psychological Science, 6, 291–298. Wetzels, R., & Wagenmakers, E.-J. (2012). A default Bayesian hypothesis test for correlations and partial correlations. Psychonomic Bulletin & Review, 19, 1057–1064. Wicherts, J. M., Borsboom, D., Kats, J., & Molenaar, D. (2006). The poor availability of psychological research data for reanalysis. American Psychologist, 61, 726–728. QUANTIFYING REPLICATION RESULTS 30 Yong, E. (2012). Nobel laureate challenges psychologists to clean up their act: Social–priming research needs “daisy chain” of replication. Nature. Zellner, A. (1986). On assessing prior distributions and Bayesian regression analysis with g–prior distributions. In P. Goel & A. Zellner (Eds.), Bayesian inference and decision techniques: Essays in honor of Bruno de Finetti (pp. 233–243). Amsterdam: North–Holland. Zellner, A., & Siow, A. (1980). Posterior odds ratios for selected regression hypotheses. In J. M. Bernardo, M. H. DeGroot, D. V. Lindley, & A. F. M. Smith (Eds.), Bayesian statistics (pp. 585–603). Valencia: University Press. Zwaan, R. A., & Zeelenberg, R. (2013). Replication attempts of important results in the study of cognition. Frontiers in Cognition. QUANTIFYING REPLICATION RESULTS 31 Appendix A Normal Approximation to the Posterior of the Original Experiment The probability density of the observed t value given a single effect √ size δ is given by the non-central t-distribution with non-centrality parameter ∆ = δ N evaluated at t = tobs , that is, p(tobs | δ) = tdf,∆ (tobs ). It follows that the posterior distribution for δ can be represented as a combination of non-central t densities evaluated at t = tobs , each density being associated with a particular value of δ. Iterative Normal Approximation In order to approximate the posterior distribution for δ we used the method discussed by Cumming and Finch (2001), who showed that a frequentist 95% confidence interval for δ can be derived from the combination of non-central t-distributions t(n−1),∆ . To compute this confidence interval we need to find upper and lower bound values for ∆ such that the corresponding non-central t-distributions give rise to the observed t value tobs with a probability of .025 in the upper and lower tail. The upper and lower bound values for ∆ need to be approximated through an iterative procedure. After the bounds for ∆ have been √ obtained, the corresponding values for the parameter of interest δ can be found by ∆/ N . We use this approximation of the confidence interval for ∆ to construct a normal approximation to the posterior distribution of ∆ with the lower bound values found for ∆ as the .025 quantile of the approximate normal distribution with a mean value equal to ∆ = tobs as found in the observed data. Finally, the posterior distribution for effect size δ √ is obtained by the transformation δ = ∆/ N . Lambda-Prime Normal Approximation Another way to derive the posterior distribution for the non-centrality parameter δ was developed by Lecoutre (1999, 2007). He showed that, given a non-informative prior which does not favor any particular value of ∆, the posterior distribution of ∆ is a combination of the z and χ2 distributions. This Lambda-prime distribution is a non-central distribution, again with df degrees of freedom, but with non-centrality tobs : p(∆ | Y ) ∼ Λ′df (tobs ), p Λ′df (tobs ) = z + tobs χ2 /df . (7) This method has the conceptual advantage that it involves a unique distribution. However, the Lambda-prime distribution is somewhat complicated and a straightforward normal approximation can be derived as follows (Lecoutre, 2007): Λ′df (tobs ) ≈ N (M, V ) M = ktobs V = 1 + t2obs − M 2       log(2) − log(df ) df + 1 df log(k) = + log Γ − log Γ . 2 2 2 (8) QUANTIFYING REPLICATION RESULTS 32 Finally, the posterior distribution for effect size δ is obtained by the transformation √ δ = ∆/ N . Closeness of Approximation Both the iterative and the lambda-prime normal approximations are close to the approximated posterior distribution for δ, even when the sample size is small.8 Figure A shows the approximated posterior density and both normal approximations for tobs = 2. Each panel shows the results for a different sample size. It is evident that with increasing sample size, the approximations become highly accurate. Appendix B The Equality-of-Effect-Size Bayes Factor Test The relation between the data and the effect sizes, p(Yrep , Yorig | δorig , δrep ) can be expressed through the t values. Bayarri and Mayoral (2002b) assumed that the t values have a non-central t-distribution around the effect size δi with a unique study-specific variance σi2 . They further assumed that these effect sizes δi are in turn normally distributed around an underlying main effect size µ. The model and priors used can be expressed through the following distributions: p ti | δi , σi ∼ N (δi (n∗i σi2 ), σi2 ), i = orig, rep δi | µ, τ ∼ N (µ, τ 2 ), i = orig, rep σi2 ∼ IG(νi /2, νi /2), i = orig, rep (9) p(µ) ∝ 1 τ 2 ∼ IG(a + 1, ak) where n∗i = (1/ni1 + 1/ni2 )−1 and the degrees of freedom νi = ni1 + ni2 − 2 , for i = orig, rep. IG indicates the Inverse gamma distribution and N the normal distribution. A Bayes factor to test the hypothesis of equality of effect sizes can be defined through the conditional posterior predictive distributions for the replicated t value trep given the original t value torig . s !! ∗ σ2 n 1 1 rep rep 2 2 2 (10) p(trep | torig , τ 2 , σorig , σrep ) = N trep | torig , n∗rep σrep + + 2τ 2 2 n∗rep n∗orig n∗orig σorig This posterior predictive distribution is evaluated at τ 2 = 0 to acquire p(Yrep , Yorig | H0 ) and integrated over τ 2 | H1 to acquire p(Yrep , Yorig | H1 ), but in addition both expressions need to be integrated over the study-specific variance σi2 : R 2 , σ 2 )p(σ 2 , σ 2 | t 2 2 p(trep | torig , τ 2 = 0, σorig orig )d(σorig , σrep ) rep orig rep , (11) B01 = R 2 2 )p(τ 2 , σ 2 , σ 2 | t 2 2 p(trep | torig , τ 2 , σo2 , σrep orig )d(τ , σorig , σrep ) orig rep This Bayes factor can be easily computed by sampling from the prior distributions of σ 2 and τ 2 and computing the average under H0 and H1 over all samples. 8 The code for these approximations is available at http://www.josineverhagen.com/?page_id=76. QUANTIFYING REPLICATION RESULTS 33 Figure A1. Accuracy of the lambda-prime and iterative approximations to the posterior for δ. In this example, t = 2. Each panel shows the result for a different sample size. Both normal approximations are relatively close, even when sample size is low.