On the belief that beliefs should change according to evidence: Implications for conspiratorial, moral, paranormal, political, religious, and science beliefs

Does one’s stance toward evidence evaluation and belief revision have relevance for actual beliefs? We investigate the role of endorsing an actively open-minded thinking style about evidence (AOT-E) on a wide range of beliefs, values, and opinions. Participants indicated the extent to which they think beliefs (Study 1) or opinions (Studies 2 and 3) ought to change according to evidence on an 8-item scale. Across three studies with 1,692 participants from two different sources (Mechanical Turk and Lucid for Academics), we find that our short AOT-E scale correlates negatively with beliefs about topics ranging from extrasensory perception, to respect for tradition, to abortion, to God; and positively with topics ranging from anthropogenic global warming to support for free speech on college campuses. More broadly, the belief that beliefs should change according to evidence was robustly associated with political liberalism, the rejection of traditional moral values, the acceptance of science, and skepticism about religious, paranormal, and conspiratorial claims. However, we also find that AOT-E is more strongly predictive for political liberals (Democrats) than conservatives (Republicans). We conclude that socio-cognitive theories of belief (both specific and general) should take into account people’s beliefs about when and how beliefs should change – that is, meta-beliefs – but that further work is required to understand how meta-beliefs about evidence interact with political ideology.


Introduction
Rational action requires evidence. Given that beliefs inform action, beliefs ought to be informed by evidence. A longstanding broad perspective on human cognition holds that reason is, at least to some extent, responsible for accurate belief formation (Baron, 2008;Kohlberg, 1969;Piaget, 1932;Stanovich, 2005). However, the human capacity to revise beliefs in the face of conflicting evidence is, charitably, imperfect. Humans are prone to motivated reasoning (Kunda, 1990), identity protective cognition (Kahan et al., 2012), confirmation bias (Nickerson, 1998), myside bias (Perkins, 2019;Stanovich, West & Toplak, 2013), naïve realism (Ross & Ward, 1996), and bias blind spots (Pronin, Lin & Ross, 2002). There is widespread disagreement about the role and consequences of the human capacity to reason.
Various analogies have been used to simplify the various broad perspectives on human thought and, although they may be oversimplifications, they illustrate the disagreement. For example, it has been argued that human reasoning is better characterized by analogy to that of lawyers than philosophers (Haidt, 2012;Haidt, 2001) -that is, the function of human reason is to form arguments to convince others, as is the goal of lawyers, and not necessarily to form accurate beliefs, as is the goal of philosophers (Mercier, 2016;Mercier & Sperber, 2011). Of course, the analogy does not imply that people only reason like lawyers or like philosophers, but rather that the typical characteristics of human cognition are more similar to one frame of thinking than the other. To simplify, some researchers have disputed the common idea that reasoning facilitates sound judgment by pointing to cases (e.g., motivated reasoning) where explicit reasoning actually hurts judgment (Kahan et al., 2012;Kahan, Peters, Dawson & Slovic, 2017).
Relatedly, given evidence that we rely heavily on a number of heuristics and biases (Kahneman, Slovic & Tversky, 1982) and that unconscious processes have an (apparently) widespread impact on our decisions (Bargh & Chartrand, 1999), a prominent perspective is that explicit reasoning and deliberation is just not very effective in the context of powerful intuitions (e.g., Bargh, 1999;Bargh, Schwader, Hailey, Dyer & Boothby, 2012;Dijksterhuis & Strick, 2016;Gigerenzer, 2007;Haidt, 2001). One famous analogy is that human cognition is like an emotional (or intuitive) dog with a rational tail (Haidt, 2001) (or, in a more recent analogy, an intuitive elephant and an analytic rider; Haidt, 2012): That is, our capacity to reason does not effectively override our intuitions and emotional impulses.
These three perspectives can be summarized simplistically in terms of three general claims about the nature of human reasoning: 1) that reasoning prototypically helps make good decisions and come to informed beliefs ("reasoning is helpful"); 2) that reasoning is prototypically ineffective, since intuition dominates human cognition ("reasoning is helpless"); and 3) that reasoning prototypically undermines sound judgment and exacerbates motivated reasoning and (for example) political polarization ("reasoning is hurtful"). Although any of the three accounts may be the best explanation for the underlying psychology behind any given belief/opinion/value, the critical question here is which accounts offers the best broad description of high-level human cognition (i.e., which has the greatest explanatory power across various beliefs/opinions/values).
Although recent work has attempted to mediate between these three broad accounts by investigating individual differences in analytic thinking (e.g., Pennycook, Fugelsang & Koehler, 2015a;Pennycook & Rand, 2019b;Pennycook, 2018), this work is vague on the specific aspects of analytic thinking that support good thinking. Here we contend that people differ in terms of their explicit stance toward whether beliefs ought to change according to evidence and that this has major consequences for what beliefs, opinions, and values that they hold. That is, some may place stronger value in changing their beliefs and taking relevant evidence into account -and thereby (for example) take a stance toward reasoning that is more akin to a philosopher -whereas others may place stronger value in maintaining constancy and the defense of prior beliefs -and thereby take a stance toward reasoning that is more akin to a lawyer. Moreover, this meta-belief may impact what sort of beliefs individuals endorse as adults, indicating that reasoning really does have an impact on intuitive beliefs. The goal of the present work is to investigate these possibilities.

Is reasoning helpful or helpless?
Dual-process theories of reasoning, which distinguish from autonomous (intuitive) processes and those that are accomplished via some form of deliberative control (De Neys, 2017;Evans & Stanovich, 2013;Pennycook, Fugelsang & Koehler, 2015b;Thompson, Prowse Turner & Pennycook, 2011), typically emphasize how controlled reasoning processes can override (sometimes) incorrect intuitive responses. Although this emphasis does not imply that reasoning and accuracy are synonymous, it does suggest that there are meaningful and important cases where such an association is present (Evans, 2012). To take a recent example, individuals who are more disposed toward thinking analytically (as indexed by the Cognitive Reflection Test; Frederick, 2005) are less likely to fall for fake news regardless of whether it is consistent or inconsistent with their political ideology (Pennycook & Rand, 2019b). There is also evidence that analytic thinking is associated with disbelief in a variety of epistemically suspect beliefs (Pennycook et al., 2015a), such as in paranormal and religious beliefs (Pennycook, Cheyne, Seli, Koehler & Fugelsang, 2012;Pennycook, Ross, Koehler & Fugelsang, 2016;Shenhav, Rand & Greene, 2012), conspiratorial ideation (Swami, Voracek, Stieger, Tran & Furnham, 2014), anti-science beliefs (and specifically rejection of evolution) (Gervais, 2015), and pseudo-profound bullshit receptivity (Pennycook, Cheyne, Barr, Koehler & Fugelsang, 2015). In addition, there is evidence that reliance on intuition is associated with traditional moral values (Pennycook, Cheyne, Barr, Koehler & Fugelsang, 2014;Royzman, Landy & Goodwin, 2014) and conservative political ideology (Jost, 2017) (but perhaps more-so with political apathy; see Pennycook & Rand, 2019a).
In contrast, a large and diverse body of evidence supports the idea that reason is perhaps overrated. For instance, intuitive heuristics are often extremely useful and, in some contexts, may actually be more accurate than reasoned reflection (Gigerenzer, 2007;Gigerenzer, Todd & ABC Research Group, 1999). Illustrative (albeit extreme) examples of this come from research on expertise (Klein, 2008), which shows that naturalistic decision making (e.g., among chess masters or firefighters) allows for very rapid yet extremely accurate choice (Kahneman & Klein, 2009). Furthermore, social psychology in the 1990's provided numerous examples of the surprising power of intuition (Bargh & Chartrand, 1999;Dijksterhuis & Strick, 2016;Haidt, 2012). Although there have been questions about the replicability of some of these effects (e.g., for so-called "social priming" [Cesario, 2014]), the strong influence of intuition on decision making is not a matter of dispute (Evans, 2008;Kahneman & Klein, 2009).

Does reasoning undermine sound judgment?
In contrast to the work just reviewed, there is also considerable evidence for motivated reasoning effects (Kunda, 1990) -that is, cases where reasoning actively hurts sound judgment and causes people to become further entrenched in what they already believe (Kahan, 2013). For example, people tend to dismiss information that is inconsistent with their political ideology (Lodge & Taber, 2005;Redlawsk, 2002;Redlawsk, Civettini & Emmerson, 2010;Strickland, Taber & Lodge, 2011) and engage in biased search for information that is supportive of their beliefs (i.e., confirmation bias; Knobloch-Westerwick, Mothes & Polavin, 2017;Nickerson, 1998). In fact, there is evidence that political polarization about contentious scientific issues (such as global warming) is actually greater among individuals who are more intelligent (Kahan et al., 2012;Kahan, Peters, Dawson & Slovic, 2017;Sarathchandra, Navin, Largent & McCright, 2018) and who report having a more actively open-minded thinking style (Kahan & Corbin, 2016; but see Baron, 2017).
A parsimonious broad account of these findings is that individuals engage analytic reasoning processes, not in the service of accuracy, but as a means to protect their identity (Kahan, 2013) or to form convincing arguments (Mercier, 2016). This perspective flips the common conception of human reasoning on its head and suggests that reasoning often makes people more unreasonable. Consistent with this account, a recent meta-analysis indicated that partisan bias effects (motivated reasoning) were equivalent across the political spectrum (Ditto et al., 2019; but see Baron & Jost, 2019). This research indicates that reasoning is typically (or, at least, frequently) used in service of justifying prior beliefs, as opposed to updating them based on the evidence presented. To investigate this issue, we will focus on the idea that individuals who are more prone to engage in reasoning are more (not less) politically polarized. Consistent evidence for increased polarization among highly reflective people would indeed indicate that motivated reasoning is to be expected; to return to an earlier analogy, that humans reason more like lawyers than philosophers.

Actively open-minded thinking
Despite research showing evidence for motivated reasoning and the power of intuitions, the previously reviewed associations between analytic thinking and various beliefs/values suggests that reasoning is nonetheless used to modify beliefs in everyday life (although other factors are of course involved in determining what people believe). That is, people who are more reflective when they are given a trick question from the Cognitive Reflection Test (CRT) have different beliefs than intuitive people. A parsimonious explanation of this is that the same people who reflect on the CRT also tend to reflect about their beliefs (i.e., they use reason to modify beliefs). Nonetheless, the disposition to engage analytic thinking is not the same as having an actively open-minded stance in general (Baron, 1985;Stanovich & West, 1997) or toward evidence in particular (Baron, 2019;Baron, Scott, Fincher & Metz, 2015). Indeed, analytic thinking may be used to both override intuitions (i.e., to modify or undermine prior beliefs) or to rationalize or bolster intuitions (i.e., to reinforce prior beliefs) (Pennycook, Fugelsang, et al., 2015b).
In the present work, we will focus instead on people's beliefs about whether beliefs and opinions should change according to evidence. Moreover, we will investigate a wide range of beliefs, values, and opinion together as a way to systematically assess the potential long-term impact of people's thinking style on what they think.
The idea that some people may not be disposed to using evidence to inform their beliefs has been broached previously. For example, people may differ in their "criteria" for belief; although some hold that evidence and scientific consensus are most important, others believe that "knowledge of the heart" should also be a central consideration (Metz, Weis-berg & Weisberg, 2018). Indeed, the actively open-minded thinking scale (AOT) was created to assess (in part) the belief that it is good to seek evidence that may conflict with intuitions (Baron, 2008;Baron et al., 2015;Baron, 1985;Stanovich & West, 1997; see also Price, Ottati, Wilson & Kim, 2015, for a measure based more on self-report) -a tendency that is associated with improved decision making over and above intelligence or cognitive ability (Stanovich & West, 2000;Stanovich & West, 1998). Moreover, much like individual differences in cognitive reflection, high AOT has been linked to skepticism about supernatural claims (Baron et al., 2015;Pennycook, Cheyne, Barr, Koehler & Fugelsang, 2014;Svedholm & Lindeman, 2013) and superstition (Sá, West, & Stanovich, 1999), indicating that the AOT scale may index some aspects of openness to evidence in belief formation and revision. Indeed, Svedholm-Häkkinen and Lindeman (2018) found a "fact resistance" factor within the broader AOT measure that consists of items that ask about beliefs about changing beliefs according to evidence. As noted by Baron (2019), it is this "flexible thinking" dimension that is most central to the concept of AOT. Shortened versions of the AOT scale have also typically focused largely on the belief revision questions (Baron et al., 2015;Haran, Ritov & Mellers, 2013), which further suggests that these items are of particular relevance for the AOT's predictive validity. Nonetheless, the broad consequences of this metabelief across a variety of domains has not yet been systematically investigated despite having major relevance for several broad theories of human cognition. We will refer to our subscale simply as actively open-minded thinking about evidence (AOT-E). The items for our scale can be found in Table  1. Our AOT-E scale is not the same as has been used in the past, although some of the items are from previous (longer) versions of the AOT. For further information on how we derived the AOT-E scale from the larger full AOT scale, see the two validation studies presented in the supplementary materials.

Current work
Is reasoning prototypically helpful, helpless, or hurtful? One possibility that has not yet been broached is that the three perspectives are primarily describing different people. That is, people have different beliefs about whether beliefs should change according to evidence ("meta-beliefs") and this has consequences for the effectiveness of their reasoning and, therefore, what types of beliefs that they hold. The goal of the present investigation is to determine whether AOT-E is correlated with as wide a variety of beliefs, values, and opinions as is feasible in a single study. If AOT-E is consequential, it should be associated with people's stances on a number of important issues. To this end, we investigated conspiratorial, moral, paranormal, political, religious, and science beliefs.

AOT-E.
We administered the AOT-E scale that is presented in Table 1. Participants responded on a scale from 1) Strongly disagree to 6) Strongly agree. The AOT-E had strong reliability ( = .87). Participants rated themselves as, on average, willing to change their beliefs according to evidence (M = 69.8, SD = 19.1 -scale ranges from 0-100).
Only 19.2% of the participants were at or below the scale midpoint (indicating a resistance to evidence).

Conspiracist ideation.
Participants completed a 15-item general conspiracy beliefs scale (Brotherton, French & Pickering, 2013). The scale included items such as "A small, secret group of people is responsible for making all major world decisions, such as going to war" ( = .97). Responses were made on the following 5-point scale: 1) Definitely not true, 2) Probably not true, 3) Not sure/cannot decide, 4) Probably true, 5) Definitely true.
Paranormal belief. Participants completed a slightly revised Paranormal Belief Scale (Pennycook, Cheyne, Seli, Koehler & Fugelsang, 2012;Tobacyk, 2004) with the religious belief items excluded ( = .95). The scale consisted of 22 items sampled from 6 categories of supernatural belief (example items in parentheses): Psi ("Mind reading is possible"), Witchcraft ("Witches do exist"), Omens of luck ("Black cats can bring bad luck"), Spiritualism ("It is possible to communicate with the dead"), Extraordinary life forms ("The Loch Ness monster of Scotland exists") and Precognition ("Astrology is a way to accurately predict the future"). Participants indicated their belief by responding on a 7-point scale from 1) Strongly disagree, to 4) Uncertain, to 7) Strongly agree.
God Skepticism. Skepticism about God was assessed using the following question: "What sort of God, if any, do you believe in?" and presenting the following options of increasing skepticism (Pennycook et al., 2012;Pennycook, Ross, et al., 2016) The following note was also presented: Microaggressions are defined as "brief, everyday exchanges that send denigrating messages to certain individuals because of their group membership." All corrlations are significant at < .001 except the last, which is n.s.

Moral values.
We used Pennycook, Cheyne, Barr, Koehler, and Fugelsang's (2014) moral values scale, which consisted of 6 care/fairness ("individualising") and 4 traditional ("binding") moral values (Graham et al., 2011). Participants were asked to rate how important the values were to their moral thinking on a 7-point scale from 1) Irrelevant to 7) Extremely Important. Care/fairness values included being kind, supporting the autonomy of others, being helpful, being fair, avoiding harm, and supporting the rights of others ( = .85). Traditional values included showing respect for traditions, being patriotic and loyal, showing respect for legitimate authority, and being pure by avoiding carnal pleasures and disgusting things ( = .80).

Political ideology.
Participants were asked to indicate their stance on social and economic issues separately on scales from 1) Very liberal, to 3) Moderate, to 5) Very conservative. Following Pennycook and Rand (2019a), we computed four political categories based on the convergence between social and economic political ideology: 1) Consistent Liberals, who are liberal/very liberal on both social and eco-nomic issues, 2) Consistent Conservatives, who are conservative/very conservative on both social and economic issues, 3) Libertarians, who are liberal/very liberal on social issues but conservative/very conservative on economic issues, and 4) Consistent Moderates who are moderate on both social and economic issues. However, because of the liberal political skew of Mechanical Turk, there were only 60 Consistent Conservatives in our sample compared to 153 Consistent Liberals.
Political opinions. We also surveyed a range of political opinions (see Table 2). Participants were asked to indicate agreement/ disagreement on a 7-point scale from 1) Strongly disagree to 7) Strongly agree. As is evident from Table 2, three of the items did not correlate particularly highly with political ideology (microaggressions, campus free speech [coded so that support of free speech was counted as conservative], and men experiencing sexism). We therefore created a Conservative Opinions scale ( = .81) using all items except for these three (all items were re-scored so that a high score indicated a more conservative opinion). Participants T 3: Science beliefs scale, Study 1. Items are labelled with (rev) if they were subsequently reverse scored (for our analyses) so that a higher score indicates a more pro-scientific belief. Raw (prior to reverse scoring) correlations with conservatism are presented below. Conservatism is the mean of the social and economic conservative political ideology questions. In every case, political conservatives held a more anti-scientific stance. Mean scores range from 0-100, with 0 indicating complete disagreement with the statement and 100 indicating complete agreement. (All correlations are significant at < .001.)

Mean (SD)
Global warming Global warming is at least partly caused by human activity and is a serious problem for the environment.

Evolution
Evolution is the best explanation so far for our origins. −.43 73.7 (31.0) Big Bang The big bang theory is, generally speaking, the best explanation we have so far for the origin of our universe. were also asked to indicate their relative trust in the government on a scale from 1) Strongly Distrust to 5) Strongly Trust (this was also uncorrelated with political conservatism, r = -.07).

Free Market Ideology. Participants completed a 5-item
Free Market Ideology measure (Heath & Gifford, 2006). The scale assesses the belief in the powers of the free market ( = .85). It includes items such as: "An economic system based on free markets unrestrained by government interference automatically works best to meet human needs." Responses were provided on a 7-point scale from 1) Strongly disagree to 7) Strongly agree.

Science beliefs.
We created a science belief scale based on various contemporary scientific issues. In particular, we selected a number of typical science-related beliefs (Table 3): evolution, anthropogenic global warming, big bang theory, old Earth, and stem cell research. We also attempted to use items that have been associated with "liberal" anti-science attitudes (Table 3): resistance to technology (reverse scored), genetically modified organism (GMO) resistance (reverse scored), vaccines as a cause of autism (reverse scored), and belief in modern medicine. However, as is evident from Table 3, political conservatives were more likely to hold the more anti-scientific stance on every single issue -even issues often associated with political liberalism. Nonetheless, consistent with prior research, there was large variability in terms of how strongly conservatism predicted anti-scientific attitudes (Rutjens, Sutton & van der Lee, 2018). Participants responded on a 7-point scale from 1) Strongly disagree to 7) Strongly agree; however, for our primary analysis, all items were scored such that a higher value meant a more pro-science belief. The full scale had good reliability, = .84. Participants were also asked to indicate their relative trust in scientists on a scale from 1) Strongly Distrust to 5) Strongly Trust.
Demographics. Participants were given a demographic questionnaire that included the following items: age, gender, and English proficiency. Social and economic political ideology were included in the demographics questionnaire.

Procedure
Participants either completed the AOT-E at the beginning of the survey or at the end (but before demographics). The presentation order did not change the pattern of results and the aggregate results will therefore be reported. Otherwise, the order of the measures was as follows: 1) conspiracist ideation, 2) paranormal belief, 3) moral values, 4) science beliefs, 5) political opinions, 6) free market ideology, 7) theism, and 8) demographics (including political ideology).

Results and Discussion
As is evident from Table 4, AOT-E was strongly associated with every other primary measure. Individuals who believe that beliefs should change according to evidence (those high in AOT-E) were: a) less likely to believe conspiratorial, paranormal, and religious (and, specifically, theistic) claims, b) less likely to hold traditional moral values but were more likely to adopt care/fairness moral values, c) less conservative in terms both economic and fiscal ideology (including free market ideology) and across a range of specific political opinions, and d) less likely to hold anti-science beliefs. Gignac and Szodorai (2016) meta-analyzed typical effect sizes across social psychology and found that correlations (r) of .10, .20, and .30 can be considered relatively small, medium, and large, respectively. Using this metric, AOT-E was a remarkably strong predictor of most factors. With the exceptions of care/fairness moral values (r = .26) and the single-item fiscal conservatism (r = −.24), every effect size was above what would be considered large based on empirical norms. The correlation with conservative opinions and pro-science beliefs, in particular (r's greater than .60), were well above the 95 th percentile (r = .45) in terms of effect size norms for individual differences research in psychology (Gignac & Szodorai, 2016). This overall pattern of result undermines the idea that reasoning is ineffective and is consistent with the general claim that reasoning has major impacts on our beliefs and values.
To further understand the scope of AOT-E's predictive validity, we also investigated the extent to which it predicted specific political opinions (Table 5). With respect to political opinions, individuals who indicate being more actively openminded about evidence held broadly liberal political views. Indeed, AOT-E was less predictive for the items that were less strongly associated with political ideology: whether men experience sexism on par with women (AOT-E was significantly associated with disagreement, r = −.24); whether microaggressions are a serious problem in educational contexts (AOT-E was slightly but non-significantly associated with disagreement, r = .07) ; and whether students should be able to block controversial speakers from giving talks at their university (AOT-E was significantly associated with disagreement, r = .25). The items most strongly associated with political conservatism were most strongly negatively associated with AOT-E.
The pattern of results for individual science belief items (Table 6) was very clear (and plainly in support of the "reasoning helps" perspective): AOT-E was associated with more agreement with scientists, regardless of whether the issue pertained to agreement with a clear scientific consensus (such as around anthropogenic global warming or the big bang) or a disagreement with an anti-scientific belief (such as that GMO's are unhealthy or that vaccines cause autism). AOT-E was also positively associated with general trust in scientists.

Study 2
The results of Study 1 indicate that AOT-E is a very strong predictor of a wide range of beliefs and opinions. There are, however, three key issues that the data from Study 1 leave unresolved. The first pertains to the perhaps implausibly large effect sizes that we found in Study 1. A recent paper by Stanovich and Toplak (2019) raised an important point that pertains to the AOT (and that applies to the AOT-E): When asked about "beliefs", some individuals may assume that the We subsequently realized that the wording of this item may be problematic: Individuals might indicate that microaggressions are a problem because people are being microaggressed, but others might indicate that they are a problem because people are (falsely, presumably) indicating that they are being microaggressed.
T 5: Correlations (Pearson r) between AOT-E and political opinion items in Study 1 (Mechanical Turk). Opinion items are scored such that a higher score corresponds with a more strongly politically conservative position and are organized in order of the strength of positive association with conservatism (see Table 2). AOT-E = Actively Open-minded Thinking about Evidence. N = 375.   Campus free speech was not significantly associated with conservatism (see Table 2). At any rate, a high score indicates opposition to the idea that "students should be able to block controversial speakers from giving talks at their university". ***indicates p < .001, **indicates p < .01, *indicates p < .05.
T 6: Correlations (Pearson r) between AOT-E and science belief items in Study 1 (Mechanical Turk). Items are scored such that a higher score corresponds with a more strongly pro-science stance and are organized in order of the strength of negative association with conservatism (see Table 3). AOT-E = Actively Open-minded Thinking about Evidence. GMO = Genetically Modified Organisms. N = 375.
question is really about religious beliefs. Indeed, Stanovich and Toplak found that the extremely high correlation between AOT and religious beliefs can be partially (but not fully) accounted for using "belief revision" items (that is, the same class of items that make up the AOT-E). Of course, this may be partly due to the possibility that AOT-E plays a major role in belief formation (as we have argued). However, it may also be the case that religious believers are particularly opposed to revising their religious beliefs, but less opposed to revising beliefs in general. Consistent with both of these T 7: A revised Actively Open-minded Thinking about Evidence (AOT-E) scale that asks about "opinions" instead of "beliefs".

Original AOT-E Revised AOT-E
A person should always consider new possibilities. A person should always consider new information. People should always take into consideration evidence that goes against their beliefs.
People should always take into consideration evidence that goes against their opinions. It is important to persevere in your beliefs even when evidence is brought to bear against them. (rev) It is important to persevere in your opinions even when evidence is brought to bear against them. (rev) Certain beliefs are just too important to abandon no matter how good a case can be made against them. (rev) Certain opinions are just too important to abandon no matter how good a case can be made against them. (rev) One should disregard evidence that conflicts with your established beliefs. (rev) One should disregard evidence that conflicts with your established opinions. (rev) Beliefs should always be revised in response to new information or evidence.
Opinions should always be revised in response to new information or evidence. No one can talk me out of something I know is right. (rev) It is possible for someone to convince me to change my mind. I believe that loyalty to one's ideals and principles is more important than "open-mindedness". (rev) I believe that loyalty to one's ideals and principles is more important than "open-mindedness". (rev) possibilities, Stanovich and Toplak found that items that used slightly different wording (which did not lead to the religious belief presumption) continued to predict religious belief, but not as strongly. In Study 2, we therefore modified the AOT-E to ask about "opinions" instead of "beliefs" (see Table 7). We also changed the wording of an additional item so that there would be an equal number of standard and reverse-coded items. Participants in Study 2 were either administered the original AOT-E or the revised AOT-E. Our goal was to ask whether the results of Study 1 are robust to variations in AOT-E scale wordings even if effect sizes vary somewhat.
A second drawback of Study 1 is that our sample came from Mechanical Turk and is therefore particularly unrepresentative of political conservatives. This is a notable drawback because the association between AOT-E and political opinions may differ depending on whether the individuals are politically liberal or conservative. As such, in our second study, we collected a sample from Lucid for Academics -a source that provides American samples that are nationally representative on age, gender, ethnicity, and geography (based on quota-matching), and that therefore provides a more even and representative split of liberals and conservatives (Coppock & Mcclellan, 2019;Pennycook & Rand, 2019a).
Third, many of the AOT-E correlates reported in Study 1 have, in previous research, been shown to correlate with performance on the Cognitive Reflection Test (CRT;Frederick, 2005;Pennycook, Fugelsang, et al., 2015a) -a measure intended to assess the broad disposition to think analytically and that also correlates with AOT (Toplak, West & Stanovich, 2011). Thus, in Study 2 we included the CRT to assess the relative predictive strength of AOT-E relative to CRT.

Participants
American participants were recruited from Lucid for Academics on April 19 th , 2019. We recruited 700 participants, who were randomly assigned to one of two conditions. In total, 751 participants began the study but 60 did not finish. We also removed individuals who responded affirmatively when asked if they responded randomly at any point during the survey (77 from the original AOT-E condition and 76 from the revised AOT-E condition). The resulting sample (N = 539, Mean age = 45.4) consisted of 251 males and 278 females, 1 transgender female, 1 transgender male, 3 trans/non-binary, 4 "not listed", and 1 who preferred not to answer.

Materials
Measures identical to Study 1. The following measures were administered as in Study 1: Conspiracist ideation, paranormal belief, God skepticism, moral values, political ideology, political opinions, free market ideology, and trust in scientists. Unlike in Study 1, all of the political opinion items were significantly associated with political ideology (Table  8). We therefore used all of the items to form the political opinions scale ( = .72). AOT-E. Participants were either administered the original or the revised AOT-E scale, as outlined in Table 7. Reliability is good for both scales (original: = .72; revised: = .74), albeit not as strong as in Study 1. Participants reported being more actively open-minded when asked about opinions (revised scale; M = 65.5, SD = 16.5) than beliefs (original scale; M = 56.7, SD = 17.2), t(537) = 6.07, SE = 1.45, p < .001. Whereas 43% of the sample were at or below the scale midpoint when asked about beliefs, only 20.4% were at or below the scale midpoint when asked about opinions. Thus, although only a minority indicated a resistance to evidence in both conditions, this was more common when asked about beliefs than opinions. This is what would be expected if the conflation of beliefs with religious beliefs was causing some individuals to indicate a resistance to evidence. Alternatively (or in addition), it is possible that people are simply more open to changing opinions (which may be issues of taste/preference) than beliefs (which may refer more to people's position on issues of apparent fact). At any rate, the revised AOT-E removed the apparent bias against religious individuals (Stanovich & Toplak, 2019).

Cognitive Reflection Test (CRT).
We used a re-worded version (Pennycook & Rand, 2019b) of the three-item CRT (Frederick, 2005). The CRT consists of words problems that cue an incorrect intuitive response and that therefore partially index one's disposition to engage in reflective rea-soning (Campitelli & Gerrans, 2014;Pennycook, Cheyne, Koehler & Fugelsang, 2016;Toplak et al., 2011). The Lucid sample had particularly low accuracy on the CRT (M = .16, SD = .28; i.e., 0.5 out of 3 correct, on average -70% of the sample got 0 out of 3). As a consequence, reliability was relatively low for the CRT ( = .64).

Religious belief.
In addition to the theism measure used in Study 1, we also included a full religious belief scale (via . For this, participants were asked to indicate their degree of belief in the following supernatural religious claims: afterlife, heaven, hell, miracles, angels, demons, soul, devil/Satan, and God. Participants responded on a 5-point scale from 1) Strongly disagree to 5) Strongly agree. The religious belief scale had excellent reliability ( = .95). Unfortunately, there was a significant amount of missing data (N = 90) for the religious belief scaleperhaps because it was the only scale that was administered using a matrix responding format (this was done because our intention was to administer the scales identically as they have been administered in past research).

Science beliefs.
We attempted to expand our science belief questionnaire by adding additional items for which political liberals might be expected to have more anti-scientific stances. Specifically, we asked about the following (in addition to the items from Study 1; see Table 9): the heritability of human intelligence, the role of genetics in success, "detox" therapies, and nuclear power. However, as is evident from Table 9, the only anti-scientific stance that was more common among political liberals was opposition to nuclear power. Nonetheless, unlike Study 1, many of the issues (6 out of 13) did not significantly correlate with political ideology. At any rate, the full scale had acceptable reliability ( = .72).
Political party. In additional to the political ideology questions that were administered in Study 1, we also asked participants to indicate which political party they most strongly affiliate with: Democrat, Republican, Independent, Other. The sample was fairly politically balanced: 37% Democrat, 31% Republican, 29% Independent, and 3% "other". We also asked them who they voted for in the 2016 Presidential Election, about favorability toward Donald Trump, and to indicate the likelihood that they would vote for Trump in the 2020 Presidential Election. These measures, along with social and economic political ideology, were included in the demographics section of the survey.
Demographics. Participants were given a demographic questionnaire that included the following items: age, gender, English proficiency, education, income, and ethnicity.
T 9: Science beliefs scale, Study 2. Items are labelled with (rev) if they were subsequently reverse scored (for our analyses) so that a higher score indicates a more pro-scientific belief. Conservatism is the mean of the social and economic conservative political ideology questions. Mean scores range from 0-100, with 0 indicating complete disagreement with the statement and 100 indicating complete agreement.

Name
Item Correlation (r) with conservatism

Mean (SD)
Global warming Global warming is at least partly caused by human activity and is a serious problem for the environment.

Big Bang
The big bang theory is, generally speaking, the best explanation we have so far for the origin of our universe.

Evolution
Evolution is the best explanation so far for our origins. −.27 * * * 57.6 (35.2) Old Earth The universe is billions of years old. Nuclear power is a safe and viable source of energy. .09 * 49.6 (29.0) ***indicates p < .001, *indicates p < .05.

Procedure
Participants either completed the AOT-E at the beginning of the survey or at the end (but before CRT and demographics). The presentation order did not change the pattern of results and the aggregate results will therefore be reported. Otherwise, the order of the following measures was randomized for each participant (unlike Study 1, which used a fixed order): 1) conspiracist ideation, 2) paranormal belief, 3) moral values, 4) science beliefs, 5) political opinions and free market ideology, and 6) religious belief and God skepticism. This block of questionnaires was followed by the CRT and, finally, demographics.

Results and Discussion
As is evident from Table 10 -and again supportive of the "reasoning helps" perspective -both versions of the AOT-E scale were significantly associated with every other primary measure. However, consistent with Stanovich and Toplak (2019), the correlation between the original AOT-E and religious belief (r = .42) was more than double the size of the correlation for the revised AOT-E (r = .20). The revised scale also had decreased correlations with traditional moral values (r's = −.37 and −.17 for original and revised, respectively) and conservative opinions (r's = −.55 and −.36 for original and revised, respectively). Nonetheless, as mentioned, the revised AOT-E was a significant predictor in every caseand, based on the norms from Gignac and Szodorai (2016), most of the correlations were medium (r = .20) to large (r = .30). Moreover, both AOT-E scales were generally more strongly correlated with the measures of interest than was CRT performance. Indeed, every measure was significantly correlated with the revised AOT-E after controlling for CRT performance (all r partial 's > .16, all p's < .015). Thus, it appears that one's mere stance toward revising beliefs according to evidence may play a role in what they believe (as adults) -a conclusion that is plainly supportive of the idea that reasoning is largely effective (for some). The pattern of correlations for the individual political opinion items was similar to Study 1 (albeit with slightly weaker effect sizes; see Table 11). Both versions of the AOT-E scale were significantly associated with liberal political stances on almost every issue, with two exceptions. The first exception, as in Study 1, was that AOT-E did not correlate with believing that microaggressions are problematic or unproblematic (in Table 11 this is coded such that a higher score indicates believing that microaggressions are unproblematic). The only notable difference between the two versions of the AOT-E (apart from the fact that the correla- tions tended to be stronger for the original than the revised version) was that a more strongly pro-free speech stance was nominally negatively correlated (r = −.11, p = .070) with the original AOT-E, but significantly positively correlated (r = .19, p = .002) with the revised AOT-E; this correlation was also positive using the original AOT-E in Study 1 (r = .25, p < .001; see Table 5). The latter correlation is notable because, in the Study 2 Lucid sample, conservatives more strongly disagreed that "students should be able to block controversial speakers from giving talks at their university" -a stance that was also associated with higher AOT-E (see also De keersmaecker, Bostyn, Hiel & Roets, 2020, for related results); this correlation in the same direction in the Study 1 MTurk sample (−.08, Table 2) but was not significant. In other words, the campus free speech item is the only case where higher AOT-E is associated with a stance (favoring free speech) that is positively (although modestly) correlated with conservative political ideology (Table 8). All other issues were in the opposite direction (or non-significant, as is the case for the microaggressions item). Finally, as with the overall measures, CRT was a weaker (and often non-significant) predictor for every item relative to either AOT-E scale. Combined with Study 1, these results indicate that a major consequence of AOT-E is for political ideology -precisely the domain where motivated reasoning is purported to dominate (but for a more direct test, see Study 3).
The results for the science beliefs questionnaire largely replicated Study 1 (Table 12). That is, every science belief item that was included in both studies -including general trust in scientists -was positively correlated with both versions of the AOT-E (with the exception of the modern medicine item, which was only marginally correlated with the original AOT-E in Study 2, r = .12, p = .060). The results for the new items that were added to Study 2 were more tepid. Although disbelief in the "detoxing the body of chemicals" item was correlated with AOT-E, this was not true for any of the other new items. If anything, having a positive stance on nuclear power (the only item positively correlated, however modestly, with political conservatism; see Table 9) was nominally (but not significantly) negatively associated with the revised AOT-E (r = −.11, p = .087). Nonetheless, 10 out of 13 items (along with general trust in scientists) were correlated with the revised AOT-E in the expected direction (see also; McPhetres & Pennycook, 2020). Thus, the results again support the contention that reasoning (on balance) facilitates pro-science judgment.

Study 3
The results of Study 2 largely reinforced what we found in Study 1: Believing that beliefs (or opinions) should change according to evidence was associated with skepticism about conspiratorial, paranormal, and religious claims. Consistent T 11: Correlations (Pearson r) between AOT-E and political opinion items in Study 2 (Lucid). Opinion items are scored such that a higher score corresponds with a more strongly politically conservative position and are organized in order of the strength of positive association with conservatism (see Table 8 with Stanovich and Toplak (2019), asking about opinions (revised AOT-E) instead of beliefs (original AOT-E) decreased (but did not wholly undermine) the correlation with religious belief -nonetheless, the revised AOT-E continued to significantly predict religious belief. Moreover, the revised AOT-E was just as successful at predicting conspiratorial and paranormal beliefs as the original AOT-E. Furthermore, as in Study 1, AOT-E was positively associated with care/fairness moral values and negatively associated with traditional moral values. Both versions of the AOT-E were also negatively correlated with political conservatism; including political ideology, free market ideology, and a wide range of conservative political opinions. The only exception was that the revised AOT-E was positively associated with support for campus free speech. Although this item was only modestly associated with political conservatism (r = .09), it is noteworthy that this is the sole issue out of the ten surveyed where the more politically conservative stance was associated with the stance that beliefs should change according to evidence (see also De keersmaecker et al., 2020). Both versions of the AOT-E were also predictive of a number of pro-science beliefs (with a few exceptions) (McPhetres & Pennycook, 2020). Overall, these results indicate that the AOT-E scale maintains strong predictive validity even if "opinions" are referenced instead of "beliefs". Although Study 2 paints a fairly clear picture in the ag-gregate, it remains unclear if AOT-E is predictive of (in particular) liberal opinions and pro-science beliefs across the political spectrum. Indeed, previous research has shown that cognitive sophistication interacts with political ideology when predicting people's stance on issues such as global warming (Kahan et al., 2012;Kahan, Peters, Dawson & Slovic, 2017;Sarathchandra, Navin, Largent & McCright, 2018). Unfortunately, because we assigned participants to two different AOT-E scales, we did not have enough power in Study 2 to effectively estimate effect sizes when separating Democrats and Republicans. Given that the original "belief" version of the AOT-E may modestly inflate some estimates of the correlation between AOT-E and a variety of issues (particularly those that have some association with religious belief), we ran a third study employing only the revised "opinion" version of the AOT-E.

Participants
American participants were recruited from Lucid for Academics on May 9 th , 2019. We recruited 1000 participants. In total, 1063 participants began the study but 103 did not Only 97 Democrats, 88 Republicans, and 75 Independents were administered the revised AOT-E in Study 2.

Materials and Procedure
The materials and procedure were identical to Study 2, with the following exceptions: 1) Participants were administered only the revised ("opinion") AOT-E; 2) the religious belief questionnaire (for which there was substantial missing data in Study 2) was changed from a matrix presentation format to the single-question format used for other measures; 3) we also changed the response options for the religious belief questionnaire to be consistent with the paranormal/political/science questionnaires (i.e., a 7-point scale); 4) we added 3 CRT items from Thomson and Oppenheimer (2016) that are relatively easier, based on past research (see https://osf.io/xqzse/ for full materials); and 5) we added a single continuous measure of Democrat-Republican preference ("Which of the following best describes your political preference?" Strongly Democratic, Democratic, Lean Democratic, Lean Republican, Republican, Strongly Republican), in addition to the party classification item used in Study 2 (which included "independent" as an option). Scale reliabilities for Democrats, Republicans, and Independents are in Table 13. Reliability was low for the free market ideology scale and (among Republicans and Independents in particular) the conservative opinion and pro-science belief scales. Notably, variability was fairly similar across the major variables for Democrats and Republicans, indicating that restriction of range is not a likely explanation for any divergences that we observe.

Results and Discussion
Our focus for Study 3 was on the extent to which AOT-E predicted the same constellation of beliefs, values, and opinion for individuals across the political spectrum. As is evident from Table 14, AOT-E was a strong predictor across the board for Democrats (paralleling the overall results from Studies 1 and 2). However, the same was not equally true for Republicans, for whom AOT-E was a significant predictor of skepticism about paranormal claims, acceptance of care/fairness as moral values, and (notably) pro-science beliefs, but no other issues. Interestingly, in contrast with the overall results, AOT-E was positively associated with economic conservatism among Republicans. Nonetheless, it is noteworthy that Republicans scored themselves lower on AOT-E (M = 61.2, SD = 15) than both Democrats (M = 65.1, SD = 15.9) and Independents (M = 64.9, SD = 16.3), t's > 2.5, p's < .015. Moreover, the correlations when considering the full sample (averaging across liberals and conservatives) paralleled the previous two studies: AOT-E was a significant predictor for every measure except economic conservatism.
We will focus here on party affiliation, but the results are similar if liberals and conservatives are separated using the political ideology measures or the continuous party identification measure.
Furthermore, as in Study 2, the CRT results tended to parallel AOT-E despite being a weaker predictor overall.
Turning now to specific political issues that formed our conservative opinions scale (Table 15), it is evident that the previously identified correlation between AOT-E and liberal political opinions (with one notable exception) was driven largely by Democrats and, in some cases, by Independents. For example, those higher in AOT-E were supportive of same-sex marriage among Democrats (r = −.31) and Independents (r = −.27), but this correlation is marginally significant among Republicans (r = −.12, p = .075). Most importantly, however, there was only a single issue where AOT-E predicted opposite opinions for Democrats and Republicans: Capital punishment. Whereas higher AOT-E was associated with more opposition to capital punishment among Democrats, it was associated with more support for capital punishment among Republicans. Thus, even though AOT-E was not as strongly predictive among Republicans as Democrats, it is clear that it is not merely driving political polarization either (lest more issues would be significantly correlated in opposite directions). Indeed, the opinion that there is room for men in feminism was associated with higher AOT-E for both Democrats and Republicans. Finally, the previously noted exceptional case where the more conservative opinion was, overall, associated with higher AOT-E -support for free speech -was driven by Republicans and Independents (i.e., AOT-E was not associated with support for free speech among Democrats). Thus, whether AOT-E predicts support or opposition for a specific issue appears to depend to some extent on what the issue is and about whom you're speaking. Nonetheless, AOT-E certainly maintained a great deal of predictive validity (contrary to the "reasoning   Items are scored such that a higher score corresponds with a more strongly pro-science stance and are organized in order of the strength and direction of association with conservatism (see Table 17). is helpless" perspective and consistent with the "reasoning is helpful" perspective) and was not associated with political polarization writ large (underming the "reasoning hurts" perspective).
The results for science-related beliefs (Table 16) parallel the pattern for conservative opinions insofar as they provide evidence against the motivated reasoning ("reasoning hurts") perspective. Specifically, AOT-E was generally associated with pro-science beliefs across the board for Democrats (with the exceptions of skepticism about detoxing and the role of genetics in success, which were not significant, and supporting nuclear power, which was negatively associated with AOT-E) and Independents (with the exceptions of genetics and supporting GMO's). Among Republicans, the most politically polarizing issues, such as global warming , big bang, and evolution (see Table 17), were not associated with AOT-E. However, AOT-E was associated with pro-science stances on several intermediate issues; namely support for stem cells, vaccines, technology, and modern medicine. Furthermore, trust in scientists was positively associated with AOT-E across the political spectrum.
Consistent with Kahan and Corbin (2016), we found a significant interaction between belief in anthropogenic global warming and political partisanship (Democrat vs. Republican) in predicting AOT-E, = .30, p < .001. However, as shown in Table 16, the coefficient for the correlation between global warming and AOT-E among Republicans is actually positive (instead of negative), which is contrary to the claim that AOT-E is associated with more skepticism about global warming among Republicans. .12 * * 48.5 (28.9) ***indicates p < .001, **indicates p < .01.

General Discussion
Although the belief that beliefs (and opinions) ought to change according to evidence is held by most people, there is meaningful variability in the strength of this conviction. The results of all three studies point to one broad, yet important, conclusion: Actively open-minded thinking about evidence (AOT-E) is, in the aggregate, a strong predictor of a wide range of beliefs, opinions, and values. This implies that individual differences in the propensity to reflect about evidence is something that people meaningfully engage in their everyday lives, which indicates that the exercise of human reasoning is, on balance, "helpful". The respective ideas that reasoning is "helpless" or "hurtful" did not find support in our data.
To summarize, AOT-E was associated with skepticism about conspiratorial, paranormal, and religious claims and agreement with a variety of scientific claims. Thus, AOT-E appears to support the rejection of epistemically suspect beliefs, thereby indicating that what people believe about whether beliefs ought to change (meta-beliefs) has an influence on what they take to be true or false about the world. AOT-E was also consistently associated with political liberalism in a variety of forms (despite being a domain for which motivated reasoning should be prominent); from having a more liberal political ideology, to adoption of more liberal moral values (specifically, rejection of traditional values and agreement with care/fairness values), to opposition to economic conservatism and free-market ideology. Furthermore, AOT-E was positively associated with a variety of specific liberal political opinions (e.g., supporting gay marriage and access to abortion) in the aggregate. This suggests that political conservatives, who tend to be more resistant to societal change (White, Kinney, Danek, Smith & Harben, 2019), may also be more resistant to intrapersonal belief change (but see below for a more nuanced interpretation).
The strong predictive validity of the AOT-E across a wide range of domains suggests that people's meta-beliefs about whether and how beliefs should change play an important role in belief formation. However, this conclusion comes with an important caveat that is in many ways as interesting as the conclusion itself. Most notably, Study 3 revealed that AOT-E is much more predictive among Democrats than among Republicans (with Independents being intermediate). This interaction is, in some cases, consistent with previous research that has been used to support the "reasoning hurts" perspective -however, as we will argue, it is not consistent with how some of these past results have been interpreted.

AOT-E among Democrats and Republicans
The AOT-E did not have the same predictive validity for conservatives as it did for liberals. To take a prominent example from Study 3, AOT-E was very strongly correlated with belief in anthropogenic climate change among Democrats (r = .43, p < .001), but there was no such (significant) correlation among Republicans (r = .09, p = .179). This parallels previous findings wherein individuals who are more cognitively sophisticated (using a variety of measures, including the CRT) are more politically polarized in terms of climate change (Kahan et al., 2012). In particular, Kahan et al. found that science literacy and numeracy was positively correlated with climate change risk attitudes among liberals (r = .08, p = .03) but negatively correlated among conservatives (r = −.12, p = .03). The favored explanation for these results is that cognitive sophistication polarizes climate change (and Nonetheless, AOT-E was significantly correlated with every primary measure in all three studies even after political ideology and demographics were statistically controlled (see supplementary materials for full analysis). The same was true for CRT performance (albeit to a lesser extent), except for care/fairness moral values and free market ideology.
We see a more direct parallel when considering the association between CRT and belief in anthropogenic global warming in Study 3: r = .19, p = .001 among Democrats; r = −.12, p = .088 among Republicans. Furthermore, Kahan and Corbin (2016) also found a significant positive correlation between views on climate change and AOT among Democrats and a null (but nominally negative) correlation among Republicans. other) attitudes because it facilitates motivated ("identityprotective") reasoning (Kahan et al., 2012;Kahan, Peters, Dawson & Slovic, 2017;Sarathchandra, Navin, Largent & McCright, 2018) -an account that has notably been applied to the AOT as a measure of general cognitive sophistication as well (Kahan & Corbin, 2016; but see Baron, 2017). This "humans-as-lawyers" motivated reasoning perspective has had a large influence on the field and in the popular press (for an overview, see Pennycook, 2018).
Motivated reasoning cannot, however, account for our broad pattern of results (with some potential exceptions). Although an interaction between political ideology and cognitive sophistication in the prediction of an attitude (such as climate change belief) is often taken as positive evidence for the motivated reasoning account, the interaction is easily understood as a consequence of the sample characteristics. For example, the interaction between political ideology and cognitive sophistication in the prediction of climate change attitudes emerges because the sample happens to have both liberals and conservatives. However, an interaction can emerge from opposing effects that are not individually significant. This is important because the central prediction of the motivated reasoning ("reasoning hurts") account is actually two separate (and opposing) main effects for political liberals and conservatives: Cognitive sophistication should be positively associated with politically congruent attitudes (e.g., climate change for liberals) and negatively associated with politically incongruent attitudes (e.g., climate change for conservatives; see Pennycook & Rand, 2019b). The results of Study 3 are plainly inconsistent with this prediction: Not only did we not find opposing main effects in the context of climate change (in fact, AOT-E is nominally positively correlated with climate change beliefs among Republicans), we do not find it for any other polarized issue either (with one exception). Specifically, there was not a single scientific issue that we included in our study that produced significant correlations with AOT-E in opposite directions for Democrats and Republicans. Furthermore, across ten explicitly political issues (such as support for police authority or opposition to abortion) there was only one case where the motivated reasoning prediction of significant opposing effects was present: Capital punishment. Even broad ideological positions such as social conservatism and free-market ideology did not produce significant opposing effects (although a second exception is present here: Economic conservatism). Thus, in almost every case, the motivated reasoning (or identity-protective cognition) account's prediction was not supported. Rather, it appears that AOT-E is simply less predictive among political conservatives than among liberals.
One potential explanation of this is that, despite arguments to the contrary (Kahan & Corbin, 2016), actively open-minded thinking about evidence is not merely a proxy for cognitive sophistication (Baron et al., 2015) -an observation supported by the divergences between AOT-E and CRT in our own data (with the former being a stronger predictor overall than the latter) as well as the data of Kahan and Corbin (Baron, 2017). Indeed, as intimated in the introduction, AOT-E is definitionally opposed to motivated reasoning: Believing that beliefs ought to change according to evidence essentially amounts to a rejection of motivated reasoning. Of course, it is possible that those who report being more actively open-minded are simply being deceptive (and potentially self-deceptive). That is, individuals who say that they are particularly willing to change their beliefs according to evidence are, in reality, the most likely to do the opposite and engage in motivated reasoning. This seems unlikely. And, at any rate, the results for the CRT -which is plainly a measure of some sort of cognitive sophistication -also do not support the motivated reasoning account. There was not a single specific issue, political or sciencebased, that produced opposing and significant correlations with CRT for Democrats relative to Republicans. Thus, the most parsimonious take-away from the present data is simply that the motivated reasoning account (a "reasoning hurts" perspective) is wrong or incomplete.
If not motivated reasoning, what then explains the finding that AOT-E is more consistently predictive for liberals than conservatives? It is potentially revealing that the depression in predictive validity for Republicans relative to Democrats was seemingly evident even for measures that did not significantly correlate with conservatism. For example, conspiracy ideation was strongly correlated with AOT-E for Democrats (r = −.32, p < .001) but less so for Republicans (although it was marginally significant, r = −.12, p = .070). This occurred even though conspiracy belief was equivalent between Democrats and Republicans (t = 1.16, p = .245). One mundane possibility is that data quality was, for whatever reason, poorer among Republicans than Democrats. Contrary to this, scale reliabilities were largely similar for both groups (Table 13). Moreover, random responding was very similar for Republicans and Democrats.
One possibility is that there are important differences between the "coalitions" that make up the Democratic and Republican parties. For example, Baron (2017) noted that the Democratic Party in the United States (and liberals in general) is made up of a more diverse group of people than is the Republican Party. Supporting this idea, variation in most of the primary measures in Studies 2 and 3 (including the AOT-E itself) is at least nominally higher among Democrats than Republicans (see Table S4 in the supplementary materials).
Yet another possibility is that there is no genuine difference between conservatives and liberals in terms of people's beliefs about how beliefs should change, but the AOT-E items are viewed through a political lens in the similar sort of way that the "belief" items in the original AOT-E were biased However, the interaction between AOT-E and a continuous measure of political partisanship (Democrat vs. Republican) in the prediction of conspiracy beliefs was only marginally significant, = .06, p = .093. against religious individuals (Stanovich & Toplak, 2019). It may be that "evidence" as a term or concept has been politicized to some extent and that AOT-E may be viewed as a liberal outlook (Krugman, 2019). Under this account, the weaker correlations among Republicans occurs because some proportion of conservatives are reporting lower AOT-E simply because they are resisting the framing or wording of the questions (or perhaps the source of the questionsfor more on insincere responding in the context of partisan bias in surveys, see Bullock & Lenz, 2019). One counter to this possibility, however, is that the CRT is also less predictive among Republicans. Republicans did no worse on the test than Democrats and presumably are not ideologically opposed to simple-seeming word problems. Although this does not rule out the politicization of evidence possibility, it does render it less likely.
Alternatively, liberals and conservatives (in the USA) may genuinely differ not only in what they believe (including meta-beliefs), but why they believe it. The AOT-E is equipped to assess one's stance toward evidence, which is apparently important among liberals (insofar as AOT-E distinguishes between what types of beliefs and opinions liberals tend to have -although, of course, other factors are surely important as well). At least based on the present correlational data, belief formation appears to be driven more by other factors for conservatives. That is, it is not simply that conservatives are less willing to change their beliefs according to evidence (although the overall difference is nonetheless evident), but rather that factors unstudied here contribute more substantially to belief formation among conservatives. One of the apparent defining features of conservatism, apart from resistance to change, is the endorsement of hierarchies (e.g., Jost, Glaser, Kruglanski & Sulloway, 2003). Perhaps part of the reason why AOT-E is less predictive among conservatives, then, is that beliefs are less intrapersonal and more interpersonal among political conservatives. That is, belief is more about social groups and, therefore, variation in beliefs among conservatives is driven more by exposure to different hierarchies and information sources (for an example, see Landrum, Lull, Akin, Hasell & Jamieson, 2017). Plainly, further research is required.

Very large effect sizes: A lesson
In a recent discussion of effect size estimates, Funder and Ozer (2019) argued that r's of .10, .20, and .30 correspond to small, medium, and large effect sizes, respectively (see also Gignac & Szodorai, 2016). They also argued that very large effect sizes (r = .40 or greater) are, in the context of psychological research, "likely to be a gross overestimates rarely found in a large sample or in a replication" (p. 1). In Study 1, AOT-E predicted multiple beliefs and opinions at a level greater than .40 (including aggregate liberal opinions and pro-science beliefs at r's =˜.60). As a meaningful counter-example to Funder and Ozer, Study 1 consisted of a large sample (N = 375) and was largely replicated with a different sample (using the original AOT-E) in Study 2 (r's were .55 and .40 for liberal opinions and pro-science beliefs, respectively). Nonetheless, consistent with Funder and Ozer's larger point, the very large effect sizes in Study 1 may be inflated for two reasons.
First, as argued by Stanovich and Toplak (2019) (who also noted the large effect sizes as a reason for skepticism), the original version of the AOT-E appears to have inflated some effect sizes because individuals may have presumed the questions to be about religious belief in particular instead of beliefs more generally. Although religious believers continue to rate themselves as more resistant to revising opinions according to evidence relative to non-believers, the "belief" wording in the original AOT-E (which was derived from earlier scales) may have inflated the correlation with religious belief and its covariates. However, one alternative possibility is that the belief items are simply more predictive overall (e.g., because they are more easily understood). A more systematic investigation of "belief" versus "opinion" wording is necessary to come to firm conclusions. Of course, the present data indicate that the AOT-E is relatively strongly predictive regardless of these small changes in wording.
Second, as discussed, we found that AOT-E was much more predictive across the board for political liberals (Democrats) than for conservatives (Republicans). This is relevant for the apparently over-estimated correlation effect sizes in Study 1 because that sample came from Mechanical Turk, which was heavily liberal-skewed. Thus, although Study 1 consisted of a large sample and produced results that were replicated in Study 2, our evidence indicates that Funder & Ozer's (2019) conclusion about very large effect sizes being likely overestimates is nonetheless accurate. However, in this case, the issue was more a matter of generalizability than replicability. The underlying lesson is the same: Very large effect sizes should be interpreted with caution.

Limitations
The principal limitation of the present study is that it is correlational and therefore not possible to establish, for example, whether AOT-E affects political opinions, vice versa, and/or some third factor affects both. Nonetheless, experimentally manipulating AOT-E and testing for a change in beliefs does not seem a prudent approach. Beliefs, opinions, and values are formed across years, and minute-long manipulations do not offer a reasonable proxy for the psychological processes It is also noteworthy that reliability for the original AOT-E was greater for the Mechanical Turk sample (.87) than the Lucid sample (.72). It is possible that the muted effect sizes in Studies 2 and 3 are also the result of lower data quality on Lucid than on Mechanical Turk (and, indeed, random responding was much more common in the former than the latter). This may also explain why the CRT was broadly less predictive in the Lucid sample than in past studies. CRT scores were, overall, quite low on Lucid.
that are of chief interest here. Rather, longitudinal studies that establish differences in AOT-E at adolescence and test for changes in beliefs over time would be a more fruitful future direction.
The generalizability of this study is limited in a number of ways. First, our samples are not precisely representative of the United States population (although Lucid is much closer than Mechanical Turk). Second, we obviously cannot generalize beyond the USA. Third, although we attempted to test as many different beliefs, values, and opinions as possible in a single survey, we have surely missed many important issues. Furthermore, it is possible that our own selection of issues was subject to our own liberal political bias.

Conclusion
Our 8-item actively open-minded thinking about evidence (AOT-E) scale was strongly predictive of a wide range of beliefs, values, and opinions. People who reported believing that beliefs and opinions should change according to evidence were less likely to be religious, less likely to hold paranormal and conspiratorial beliefs, more likely to believe in a variety of scientific claims, and were more political liberal (in terms of overall ideology, partisan affiliation, moral values, and a variety of specific political opinions). Moreover, the effect sizes for these correlations was often large or very large, based on established norms (Funder & Ozer, 2019;Gignac & Szodorai, 2016). The size and diversity of AOT-E correlates strongly supports one major, if broad, conclusion: Socio-cognitive theories of belief (both specific and general) should take into account what people believe about when and how beliefs and opinions should change (i.e., meta-beliefs). That is, we should not assume that evidence is equally important for everyone. Furthermore, our findings provide clear support for the perspective that reasoning facilitates sound judgment, thereby undermining the idea that intuition commonly dominates reasoning. We also found essentially no support for motivated reasoning. Regardless future work is required to more clearly delineate why AOT-E is more predictive for political liberals than conservatives.