Credibility in Social and Personality Psychology

Hits and Misses in the Last Decade of Open Science: Researchers From Different Subfields and Career Stages Offer Personal Reflections and Suggestions

Jennifer Beer*1, Paul Eastwick2, Jin X. Goh3

Social Psychological Bulletin, 2023, Vol. 18, Article e9681, https://doi.org/10.32872/spb.9681

Received: 2022-06-14. Accepted: 2022-12-11. Published (VoR): 2023-11-17.

Handling Editors: Simine Vazire, Melbourne School of Psychological Sciences, University of Melbourne, Melbourne, Australia; Brian Nosek, University of Virginia, Charlottesville, VA, USA

*Corresponding author at: Department of Psychology, 108 E Dean Keeton Street, Stop A8000, Austin, TX, 78712. E-mail: beerutexas@gmail.com

Related: This article is part of the SPB Special Topic "Is Psychology Self-Correcting? Reflections on the Credibility Revolution in Social and Personality Psychology", Guest Editors: Simine Vazire & Brian Nosek, Social Psychological Bulletin, 18, https://doi.org/10.32872/spb.v18

This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

The success of Open Science in addressing the replication crisis and restoring credibility in psychology can be understood more completely by examining the successes and challenges of adapting the recommended best practices by researchers at different types of institutions, different career stages, and from different subfields within social and personality psychology. In this article, we offer personal reflections about the impact and future of the Open Science movement in a conversational form between three researchers at varying career stages who focus on different subdisciplines (relationship science, diversity science, and social neuroscience and social cognition) and work at universities that place a different emphasis on research (relative to teaching and service). We see many successes of the open science movement, but we also note that implementation has trailed behind its full potential because (a) the incentive structures of our existing rigid system remain misaligned with some open-science goals, and (b) some open science solutions were designed by researchers with certain types of scientific practices in mind. We all feel encouraged by the focus on larger samples, greater data sharing, and pre-registration both for experimental design and analytical decisions. However, there are areas that need attention. Our perspective is that the open science movement has not been as strong of an ally as it could be to another goal of psychological science: increasing diversity, equity, and inclusion. Additionally, more careful consideration of how to facilitate data sharing and pre-registration is needed and may necessitate a shift in the incentive structure of our field.

Keywords: open science, transparency, replication, pre-registration, diversity

Highlights

  • A conversation between researchers at different career stages and at different types of research institutions identifies new directions to fulfill the potential of Open Science goals in Psychological research.

  • New pre-registration guidelines are provided and discussed in the framework of three different goals for promoting transparency and replication.

  • We call for transparency and robust empiricism to be favored over easy routes to assign awards to individual scientists.

  • We note the ways in which the Open Science movement in Psychology could be an even stronger ally on issues of diversity, equity, and inclusion.

Is there a reason to be optimistic that psychology is doing all it can to correct questionable research practices, bolster replicability, and regain credibility? The answer to this question depends on whom you ask. While many high-profile researchers debate the utility of various best practices, less attention has been paid to systematically considering how each practice affects researchers from different subdisciplines, career stages, identities, and institutional contexts where their research is conducted. In this article, we give insight into how a recently-held conversation on the impact and future of the Open Science movement might play out between three early- and later-stage career researchers who focus on various subdisciplines (diversity science, relationship science, and social neuroscience and social cognition) and conduct their research at universities that place differing levels of emphasis on research.

The genesis of this article was as follows: Dr. Beer first broached this topic with Dr. Eastwick, and then they both invited Dr. Goh to add some additional perspective. All three authors work in academia in the United States, and they agreed that their collective expertise put them in a position to pose and answer questions at the intersection of Open Science and (a) Diversity, Equity, and Inclusion (DEI), (b) data sharing, and (c) preregistration. In order to ensure that each researcher maintains their voice in this collaborative conversation, the article indicates sections that represent thoughts from specific authors (e.g., avoiding the need to agree on an opinion that caters to the median of the group: Forscher et al., 2020).

Dr. Jin Goh received his PhD in 2017 and is an assistant professor at Colby College, a small liberal arts college; he will reflect on the changes in his subfield of diversity science, since his participation as an undergraduate student in the research consortium behind a massive replication effort of psychological findings (Open Science Research Consortium, 2015). Dr. Goh’s perspectives are shaped by his relatively early career stage and his identity as a gay Asian American man. While his race and sexual orientation may at times marginalize him within social-personality psychology, his gender and able-body privilege him to navigate and speak up comfortably in the field. Within diversity science, he uses both experimental and correlational designs to study how identities and institutions influence intergroup perceptions.

Dr. Paul Eastwick received his PhD in 2009 and is a professor at the University of California, Davis; he will reflect on the broader context of changes in the field of relationship science. Dr. Eastwick’s views of the Open Science movement were shaped by the fact that he was trained in both experimental designs and the use of multilevel modeling in dyadic and longitudinal datasets; these different knowledge bases led him to be attuned to the way that solutions might work well for one context and poorly for another. Also, at about the time that the Open Science movement was gaining prominence in Psychology around 2014, he was (nearly) tenured at a research-focused institution, so he had the luxury of having the time to adapt to the new norms while also being in a position to critique the movement—and the backlash to the movement—without incurring much personal risk.

Dr. Jennifer Beer received her PhD in 2002 and is a professor at the University of Texas at Austin; she will reflect on the progress made in both psychology and neuroscience since she participated in the same research consortium as Dr. Goh (e.g., Open Science Research Consortium, 2015). Dr. Beer’s perspective has been shaped by being a member of an underrepresented group in one of her fields, working with convenience and non-convenience samples using behavioral manipulation and neuroimaging techniques, as well as training and working at research-focused, public universities. All three authors were trained and have worked in the United States and are therefore mostly familiar with the open science movement in that context.

Our stances and opinions range from optimistic pessimism to pessimistic optimism. From all of our vantage points, we would not expect every problem to have been solved by now. And there have been many successes in the shift toward practices that are aimed at reducing questionable research practices and bolstering replicability. However, one emergent theme is that the practice(s) which are most impactful can differ by subfield. For example, the call for larger samples is great in principle but is not easily attainable by many researchers. And some areas have seen the greatest benefits from the practice of pre-registration while others have made more strides because of the recommendation to facilitate data sharing.

We note areas of convergence and divergence about the ways in which the changes in best practices have been successful and where they could use more development. We begin by describing the intersection of the open science movement and DEI issues. We next discuss the successes and failures around the shift toward greater data availability and pre-registration. We conclude with six specific recommendations to ensure that the field’s ability to correct itself includes all interested researchers and knocks down the current infrastructure which paradoxically impedes scientific progress.

How Do You Think Open Science and Diversity, Equity, and Inclusivity (DEI) Intersect?

(From PE) There is considerable potential for the open science movement to impact diversity, equity, and inclusion (DEI) in a positive way, but some goals of the movement are more obviously connected to DEI efforts than others. Consider that around 2011–2012, part of the original impetus for the open science movement was rooted in frustrations expressed by some scholars directed at other scholars whose work failed to replicate (see Nelson et al., 2018, for examples). If the movement were to revolve around grievances like these—the desire for some scholars to correct the sins of the past through exposé and criticism—such practices might improve the credibility of the existing scientific record, but it is not clear that such practices would help DEI efforts. Indeed, they might harm DEI by maintaining a climate of competition and reinforcing the notion that there are scientific “winners” and “losers” (Murphy et al., 2020).

An alternative framework is to view open science as synonymous with inclusive science (Ledgerwood et al., 2022). According to this view, the ultimate goal of the open science movement would be (a) to enable people of all backgrounds/levels of seniority to be full participants in the scientific process, (b) to reduce the advantages bestowed upon insiders and senior scholars in academic culture, and (c) to make the products of science (e.g., papers, data, code) more transparent and available for all consumers. Many of the organizations and structures that have grown out of the open science movement (e.g., the Society for the Improvement of Psychological Science, the Facebook group “PsychMAP,” The Center for Open Science) espouse values that are closer to the goals of inclusive science than the goals of a replicability-bolstering science. Nevertheless, there are likely to be continuing tensions when these goals come into conflict, such as when senior (usually male) scholars deride the presumed replicability of the work of junior scholars on unmoderated platforms like Twitter (Whitaker & Guest, 2020).

(From JG) I entered graduate school in the summer of 2012, which I spent excitedly coding articles for the Open Science Research Consortium replicability project. Discussing Bem (2011) in my undergraduate lab invigorated me to enter social psychology with an eye toward open science. I soon discovered my zeal was not widely shared in graduate school; at times I butted heads with senior faculty and graduate students. Reflecting back on the early days of the Open Science movement, my vantage point then is different from PE and even myself right now. As a graduate student then, it felt as though I was watching metaphorical giants hurling rocks (or words) at one another and everyone else had to avoid becoming casualties in their fights (or debates). Conferences, social media, and departmental meetings felt like dangerous waters because it was difficult to know who might be offended by your opinions on open science. I could not help but admire the graduate students who were willing to challenge established figures in the field.

So how did these debates impact DEI within the Open Science movement? Researchers (primarily tenured, White, and/or male researchers) at “R1” institutions dominate academia and its various leadership positions such as search committee chairs, journal editors, and professional society leaders (Buchanan et al., 2021; Roberts et al., 2020; Syed & Kathawalla, 2022). As such, their perspectives on open science were likely skewed by their experiences and positions. Yet being in positions of power, their skewed perspectives shaped the norms and journal practices that the rest of us had to follow. At the time, social-personality psychology, and the discussion of its future, was seemingly reserved for a particular demographic and everyone else was a spectator.

This situation only worsened when researchers (often from marginalized backgrounds) tried to discuss the role of DEI and received criticism for it. Some argued that open science was independent of biases because scholars were merely discussing ideas and practices and such debates should be divorced from any researcher’s identity and position. “Why bring up race or gender,” some wondered. From my perspective, this colorblind approach negated careful considerations of DEI in the early days of the movement (Ledgerwood, 2017; Srivastava, 2014).

The early Open Science movement championed itself as a free discussion of ideas and practices that all could participate, and in many ways, that is true and it is increasingly normative to discuss DEI. However, I could not help but wonder who is socialized to speak freely and comfortably of their ideas without worrying about the precariousness of their status within academia and society broadly?

(From JB) I agree with all of the points that Dr. Goh and Dr. Eastwick raise. I will also say that I have been disheartened by Twitter arguments and “behind the scenes” failures to support open science which have affected trainees’ interest in continuing in science for their career. One theme I’ve been struck by after recent conversations with students is a feeling of “doing everything right” in terms of adopting pre-registration to be transparent yet still end up being accused by editors or reviewers of QRPs which sometimes leads to the feeling that the field does not really care about improving science. Instead, we may have just exchanged one set of bullies or gatekeepers for another. And they want nothing to do with science as a result. In light of feelings like this, I have to conclude that despite some benefits, it is a huge shame if open science is driving some people away from the scientific community.

How Easy and Effective Has It Been to Adopt Data Sharing Practices Within Your Subfield?

The call for larger samples has been met with new infrastructure to facilitate collaboration (from JG).

We are all in agreement that the renewed emphasis on large samples is better for our science in terms of statistical power and effect size estimation. However, the emphasis on large samples has the potential to exclude researchers working in contexts with relatively less research support but has been met with new infrastructure to facilitate collaboration among scholars.

Instead of basing conclusions on 50 people for a factorial ANOVA, we are now seeing 500+ people in multi-study articles to achieve statistical power (Giner-Sorolla, 2018). Even in short journal formats, a perusal of published articles in 2022 showed that the majority of articles had more than one study, with total sample sizes that often exceed 500. These improvements have permitted greater statistical power and more accurate inferences, but there are limits and shortcomings. First, we are becoming overly reliant on convenient and cheap (exploitive) sampling via MTurk and other online recruitment platforms (Buhrmester et al., 2018). Due to the online nature of such studies, we may be moving farther away from rich behavioral studies that distinguish social-personality psychology from other subfields (Doliński, 2018). As an early career faculty, tenure expectations demand publications in top journals that often require multi-study packages. Many researchers (myself included) would rather rely on convenience sampling than time-intensive studies that may not pay off. The convenience of online sampling as well as the demand for large samples are pushing us to become even more of a “science of self-reports and finger movements” (Baumeister et al., 2007).

Second, increasing sample size and study numbers take time and resources that may disadvantage researchers who are not at “R1” research-intensive universities. Those at smaller teaching and non-R1 institutions tend to have a higher teaching load and do not always have graduate students or postdocs to help them with research. Subject pools also tend to be smaller at undergraduate-only institutions. As a professor at a small teaching college, I therefore have limited time and resources for research. The stakes are higher for me to spend time and money in a way that can yield the best result. This practice creates disproportionate demand depending on career stage and institutional resources.

Third, I fear diversity science is continually centering majority group members’ (e.g., White, cis, straight, male, or US American) perceptions and experiences because these participants are easier to recruit. Even with online platforms, it is still difficult to recruit minorities; people may even lie about their identities (Chandler & Paolacci, 2017). These pressures can lead to a greater emphasis on the majority group members’ perspectives.

Nonetheless, this shift also led to creative solutions, often initiated by early career researchers and those at non-R1 institutions. For example, new opportunities arose that enabled unacquainted researchers to collaborate to produce high-quality publications. In particular, the Psychological Science Accelerator (PSA; Moshontz et al., 2018) enables researchers from around the world to participate in an impactful project by contributing a minimal number of participants. Because of the sheer size and breadth of the PSA network, people from small colleges or under-resourced universities worldwide could all earn authorship in top publications like Nature Human Behavior (Jones et al., 2021).

In social neuroscience, enhanced data availability has had small effects on transparency (from JB).

Long before concerns about transparency and replicability in fields like social psychology, researchers who used neuroimaging had already shifted toward making data publicly accessible. In fact, prominent journals such as the Journal of Cognitive Neuroscience went through a period where they required researchers to put their data in a public repository (i.e., 2000–2006: Van Horn et al., 2004). And a common format (e.g., Brain Imaging Data Structure [BIDS]) and many data repositories are now available and specifically aimed at cataloguing neuroimaging data sets (e.g., OpenNeuro, 1000 Functional Connectomes Project).

However, to my eye, the enhanced accessibility to data sharing in neuroscience has not, by itself, helped transparency much for a number of reasons. First, there are a wide range of acceptable approaches to conducting the many pre-processing steps which occur before GLM or other analyses. For example, one project challenged 70 research teams to analyze the same data set and essentially found that there were as many different analytical approaches as there were teams (Botvinik-Nezer et al., 2020). Even using very similar pre-processing and analytical procedures but different software can lead to slightly different results (Bowring et al., 2019, 2021). Therefore, I would argue that while neuroimaging researchers were perhaps doing a better job of sharing data, those efforts alone were not enough to strongly combat issues of transparency or bolster conclusions about replicability.

The call for data sharing has been a clear success in relationship science (from PE).

Relationship science is a subfield that has benefitted from the emphasis on making the data from published papers broadly accessible. After overcoming a few initial challenges, relationship scientists have adjusted to these new norms and launched initiatives that put these datasets to good use.

Ten years ago, it became clear that the existing APA data sharing guidelines were insufficient. The rules, of course, were clear: “After research results are published, psychologists do not withhold the data on which their conclusions are based from other competent professionals…” (American Psychological Association, 2010, 8.14). Yet several studies revealed that only 20–40% of scholars provided their raw data to other competent professionals when asked (Vanpaemel et al., 2015; Wicherts et al., 2006). The overall unavailability of raw data meant that errors remained uncorrected, meta-analyses remained incomplete, and fraud remained undetected (Simonsohn, 2013).

The open science movement highlighted this problem, and it offered guidance in the form of both bottom-up and top-down solutions. On the bottom-up side, the open science movement cultivated norms of data sharing and transparency: Advocates stressed how scientists are more likely to believe conclusions if they can have access to the raw data themselves to check assumptions or run additional analyses. On the top-down side, many journals began to ask authors to make their data available during the publication process (or to explain why they could not). As a consequence of this combination of forces, datasets in psychological science were becoming more widely available by the middle of the last decade (Kidwell et al., 2016).

Like many of the open science reforms in psychology, data sharing guidelines were originally developed with modest research projects in mind: a paradigmatic example might consist of a few hundred participants and a handful of variables arranged in a single spreadsheet. But a relationship-science dataset could be quite large, involving hundreds of participants completing thousands of questionnaire items over months or years. It might involve many spreadsheets that would need to be linked together in complex ways, and it might also involve questions about sensitive topics (e.g., violence, infidelity, or intentions to divorce).

As guidelines at some major journals (e.g., Personality and Social Psychology Bulletin, Psychological Science) were beginning to change around 2014, relationship scientists needed to figure out exactly what these reforms meant for the datasets they tended to collect. When scholars elect to—or are required to—share such a dataset, what exactly do they share? Must they share all the data from the project, or something reduced? If the dataset has been reduced in some way, what can be cut without sacrificing the transparency goal? Many intelligent, well-intentioned scholars were puzzled by these questions, especially those who were not active on social media or who were not paying close attention to “the crisis.”

These issues prompted my contributions to an article that discussed how transparency solutions could and should be tailored to specific research areas (Finkel et al., 2015). This article channeled two persistent questions about the consequences of sharing the kinds of large datasets collected by relationships researchers. The first was whether the sharing of large data files might mean that researchers could be “scooped” by their own data, thereby disincentivizing future labor-intensive data collection efforts. This concern has proven to be misplaced, mitigated by the clear consensus that open data sharing policies cover the columns of data that can reproduce the published article. In relationship science, the full dataset could generate many independent publications, but this reduced dataset would rarely (if ever) generate a second publication on its own.

The second concern was that public data sharing would reveal too much about participants—to the world, and (especially) to their romantic partners. That is, a publicly available dataset might allow a motivated participant to find their own “row” and then, with a little bit of sleuthing, find data provided by their romantic partner. These issues took time to resolve, but there are indeed solutions (Joel et al., 2018). For example, data can often be hidden from participants themselves by posting aggregate scores (i.e., no item level or demographic data are posted), and data can be stored in a secure repository that requires institutional access (e.g., UK Datashare) and/or an application process (e.g., ICPSR). In some cases, id variables can be stripped out of the dataset without compromising the analysis, which makes finding one’s partner nearly impossible. In summary, it may take some extra work for relationships researchers to post their data, but these innovations mean that relationship scientists can and should participate in the new data sharing norms.

Critically, there have also been opportunities for close relationships researchers as they have made the transition to greater data sharing. Probably the most important development stemmed from the realization that relationships researchers possess a great deal of unanalyzed data. In practice, a typical close relationships dataset tends to generate only a small handful of papers before languishing as the investigator embarks on other projects. Furthermore, there has historically been no way to ascertain exactly how the measures from different datasets overlap, nor has there been an infrastructure to connect scholars with similar datasets who might benefit from combining their data.

Dr. Sara Algoe addressed these issues by establishing the Love Consortium in 2019. One of the primary goals of the Love Consortium was to inspire greater data sharing among close relationships researchers. After verifying that many close-relationships researchers had datasets that they were interested in making broadly available, Dr. Algoe created a repository through the UNC Dataverse. Nearly 100 datasets have been catalogued there, and the Love Consortium has awarded grants which foster collaborations on these datasets. These grants have already inspired many multi-study investigations that would never have happened otherwise. In short, the Love Consortium is a prime example of how norms toward greater data sharing and openness create scientific progress by dismantling inter-laboratory barriers and fostering a culture of collaboration.

How Has Preregistration Been Successful and Challenging?

The focus on pre-registering design, data acquisition, and analyses through pre-registration and Registered Reports has benefited subfields which were not originally the source of concerns about transparency.

(From JB). It has been exciting to see the effect of the open science movement on neighboring subfields such as social and cognitive neuroscience. My perspective is that the recommendation to pre-register analyses has been most beneficial to subfields involving neuroimaging because earlier efforts in those fields to facilitate data sharing were not especially successful to increase transparency or help with efforts to understand replicability.

The additional adoption of best practices like pre-registration and Registered Reports in fields like social and cognitive neuroscience has been a successful way in which Open Science has benefitted fields beyond those it was originally focused on improving (e.g., social psychology). In addition to access to data, it is equally important to understand which decisions, hypotheses, and analyses were planned in advance particularly in the case of neuroimaging analyses which includes a number of decisions that occur before analyses are even conducted. Additionally, the specificity of hypotheses can vary greatly. Currently, there are excellent templates available for researchers who want to pre-register their experimental design, pre-processing decisions, and analyses planned for hypothesis-testing when using techniques such as fMRI (e.g., Beyer at al., 2021). Many neuroscience-focused journals now accept Registered Reports (e.g., Cortex, Development Cognitive Neuroscience, Frontiers in Neuroscience, Nature Communications) whereby reviewers consider manuscripts on the basis of the strength of their methods and a priori analytical plan rather than the significance of the results.

The effort to replicate findings from articles randomly selected from prominent psychology journals (Open Science Research Consortium, 2015) illustrates why it is helpful to know precisely what authors were hypothesizing in terms of brain activation. One selected study had found that frontal lobe activation was associated with accepting unfair offers in a financial game (Tabibnia et al., 2008). The replication study was higher powered and also found that frontal lobe activation was associated with accepting unfair offers in that financial game; however, the precise location in the frontal lobes was different than the location found in the earlier study. There was a lot of discussion about whether the replication attempt should be considered successful and it essentially came down to whether we were considering the achievement of replication to be the confirmation of the hypothesis stated in the paper versus the precise coordinates of the activation in the results (i.e., the hypothesis was about a broad range of the frontal lobes rather than restricted to the specific coordinates that showed significant activation). The replication team favored interpreting the success of the replication as a function of the stated hypothesis. The hypothesis was admittedly about a broad area of the brain but they felt their job as the steward of the replication effort was to follow the published information to the letter rather than “reading the authors’ minds” as to what they may have meant. Others felt that the replication was only successful if the activation in the second study overlapped with activation found in the original study regardless of whether it overlapped with the stated hypothesized region outside of the site that showed significant activation. If pre-registration had been a best practice at the time of the original study, it may have assuaged concerns about whether the broad hypothesis stated in the article was indeed an a priori hypothesis and that specific coordinates or even specific brain hemispheres were not the intention.

Nevertheless, there are still challenges to be overcome. Support for pre-registration and replication often fails “behind the scenes.” All too often, researchers pre-register their plans and then reviewers either do not read the pre-registration or evaluate and comment on the manuscript as though it does not exist. For example, some researchers are still being told that their manuscript is not worthy of publication on the basis of results that do not show statistical significance despite pre-registration of experimental design and analyses that reviewers agree are quite strong. Or reviewers may ask whether researchers “peeked” at their data despite timestamps showing the data collection proceeded as stated in the pre-registration. Similarly, the adoption of Registered Reports as a general submission format has been very slow and often limited to very specific cases (i.e., replications of findings that must be justified as fundamentally important to the field) or relegated to journals that are not highly valued for researchers seeking jobs or tenure. The failure to adopt the Registered Report format for all higher-tier journals and for a wider range of research questions—and the rejection of manuscripts on the basis of non-significant results in the context of pre-registered robust methods—are not aligned with the promise of improving the transparency and replicability of science. There needs to be accountability in those spaces for following through on the adoption of creating more transparency through pre-registration and Registered Reports.

A New Approach is Needed to Realize the Benefit of Pre-Registration for Relationship Science

(From PE) Whereas pre-registration may be one of the great successes in subfields drawing on neuroscience, my view is that relationship science has a long way to go. Preregistration has not yet been widely adopted by close relationships researchers, especially those who conduct complex analyses with large, preexisting datasets. Finkel et al. (2015) tried to introduce close relationships researchers to the concept of preregistration. But unfortunately, the preregistration portion of this article seems not to have had a major impact. Following the conventions of the time, the term “preregistration” in that article referred to a wide variety of distinct practices and goals (e.g., controlling Type-I error, limiting p-hacking, combatting publication bias, and so forth). Honestly, from my current perspective in 2022, the whole section reads as a bit of a tangled mess. Here, I will explain what I believe has gone wrong with respect to the uptake of preregistration by close relationships scholars, and later on in this article I will try a different approach to encourage it.

Just as the question “Who was our discipline designed for?” can help illuminate hidden structural biases that cater to some constituencies rather than others (Ledgerwood et al., 2022), the question “Who was preregistration designed for?” can help illuminate the researcher-biases that caused the term “preregistration” to take on a wide array of goals and practices. Circa 2011, open science advocates pitched preregistration as a solution to the problem of undisclosed data analytic flexibility in the context of a certain kind of study (Simmons et al., 2011; Wagenmakers et al., 2012). Such a study would likely have been testing a social, social-cognitive, or judgment and decision-making research question, it would consist of a straightforward experimental design (e.g., a comparison between condition means), and it would require fairly simple statistics (e.g., a t-test). Given that collaborative efforts at that time failed to replicate those sorts of studies (e.g., Many Labs 1; Klein et al., 2014), it made sense that early pre-registration approaches (e.g., templates like AsPredicted.org) would be designed for such studies.

As a consequence, “preregistration” among psychologists came to be synonymous with practices like: recording a directional prediction for the purposes of falsifying a hypothesis, describing a priori data collection and data analytic decisions to limit “researcher degrees of freedom” and maintain a Type-I error rate of α = .05, and transparently recording which decisions were made before rather than after seeing the data (Simmons et al., 2017). These practices are straightforward when planning a simple two-condition study. We’ll call this form of preregistration basic experimental preregistration.

But for a close relationships researcher working with a preexisting dataset, it is often unclear how one should engage in these practices. First of all, predictions are often more complex than “group A will be higher than group B” on a single dependent variable. Also, the researcher might already know the dataset (e.g., how strongly certain variables tend to correlate), and the number of possible combinations of variables could be very large. Furthermore, there may not be a single gold-standard analytic approach, but rather dozens of reasonable analytic options, as in the social neuroscience examples that Dr. Beer mentions. For such datasets, basic experimental preregistration is (at best) extremely difficult or (at worst) epistemologically nonsensical.

A clear illustration of these challenges comes from a 2021 special issue of the Journals of Gerontology: Psychological Science. This special issue contained a series of papers on personality development that used preregistration with preexisting data, as would also be common in the close relationships field. In their introductory remarks, Lucas and Donnellan (2021) describe the challenges that arose across the special issue papers, including: the difficulty of anticipating all data processing decisions, lack of clarity about which results would receive higher vs. lower priority when interpreting the findings, and the fact that some authors had existing familiarity with the datasets. Lucas and Donnellan describe how some papers ultimately could not be included in the special issue because the large number of unanticipated issues mitigated the value of the original preregistration. In summary, preregistration is a mess in the field of close relationships, because the concept of preregistration—and all the practices it implies—were not designed with close relationships research (and similar research that uses complex data) in mind.

What Changes Would We Like to See to Help Fulfill the Potential of the Open Science Movement to Improve Psychological Science?

The three of us saw various promises and pitfalls from the past Open Science Movement in the early 2010s, but not all promises have been fulfilled and not all pitfalls have been addressed. Here, we each offer two suggestions on how we can continue to fulfill the potential of the Open Science Movement.

1. We would like to see: Continual consideration of what and who counts in open science (From JG). I am tentatively optimistic about the future of a more open and diverse science in psychology. There is active reconsideration of some practices to be more inclusive (Ledgerwood et al., 2022; Syed & Kathawalla, 2022). Building on this momentum, I hope open science continues to be reflective and become more inclusive in its meaning and participants. Open science practices have made many aspects of research accessible to those who would not normally be able to do science. But we should always be cognizant that academic cultures can preserve the status quo. Although we may consider our movements, departments, and societies to be diverse and inclusive, we risk mythologizing racial progress (Richeson, 2020). The reality is that marginalized researchers continue to be marginalized in academia (Matias et al., 2021). Creating a truly open and inclusive science requires us to actively consider our roles and positions in perpetuating or disrupting the status quo. But we cannot get there unless we reckon with the lack of diversity and inclusivity within our movements, practices, and long-held beliefs.

2. We would like to see: Changes in the incentive structure which is currently designed to undermine scientific progress and transparency (from JB). Although this may be a bit pessimistic, I don’t see the field truly getting all of the benefits of the best practices suggested to increase transparency and replicability unless the field starts centering the incentives around these practices. Right now, promotion and awards tend to be based on the contribution of an individual whereas the best science is most likely to come from teams with various expertise and, as mentioned earlier, potentially more access to collect the types of large samples or complex data which are strong building blocks of scientific knowledge (e.g., Forscher et al., 2020). In promotion and award committees, we should be asking how researchers facilitated collaborative science, inclusivity, and used transparency measures instead of simply focusing on five peer-reviewed publications to illustrate their individual contribution to science. When transparency and inclusivity considerations are incorporated at all, they tend to follow a long string of questions emphasizing the individual and are often discussed as an afterthought. Similarly, award speeches often never mention the person’s participation in transparency best practices or inclusivity efforts. Or a separate award is given as though somehow the awardee’s scientific contribution is another matter altogether.

3. We would like to see: Improvements in the motivation and opportunity to publish null or inconsistent findings (From JG). Researchers should be more willing to write up their null or inconsistent findings. Likewise, reviewers and editors should be more willing to accept such research in our journals. Null or inconsistent findings are ambiguous and difficult to interpret. However, with proper contextualization and calibration, such findings are important to our disciplines. Beyond p-values, inclusion and interpretation of other statistical information such as effect sizes and confidence intervals can offer us a richer understanding of null results.

Writing about null results requires greater sensitivity within diversity science. For instance, if we do not find majority group members demonstrating biases against a minority group, does this mean our methods were flawed (meaning it likely would not be publishable) or that times have changed and prejudice is over (unlikely given the prevalence of systemic inequality)? What happens when an expensive and time-consuming intervention field study did not yield significance? Researchers have also increasingly relied on “mini” meta-analysis to uncover a potential significant effect across a series of null findings (Goh et al., 2016). Although this meta-analytic approach offers more statistical power, I feel uncertain about the blind pursuit of p < .05. As Rosnow and Rosenthal (1989) said, “surely God loves the .06 nearly as much as the .05.”

Publishing null or inconsistent findings within diversity science is important to understand the full scope of the field, and this can only be achieved when journals are willing to publish less-than-perfect findings. For instance, Lai et al. (2016) found that most interventions could not reduce implicit biases after 48 hours, demonstrating the short lifespan of interventions and tenacity of implicit biases. Duker et al. (2022) found two common emotion regulation strategies to be ineffective in coping with gender discrimination, allowing them to consider other potential strategies that are more effective in subsequent studies. Goh et al. (2019) found that straight people did not behave differently whether their sexual minority interaction partners concealed or did not conceal their sexual orientations; the findings were carefully contextualized, Bayesian analyses were used to support the null, and the sexual minority participants’ perspectives were centered rather than over-claiming that there are no longer biases. These null results allow us to understand the full spectrum of human behaviors.

4. We would like to see: A new mindset for the pre-registration of complex data sets such as those used in relationship science (From PE). For researchers who use large and complex data sets such as those in relationship science, much more guidance about pre-registration is needed. Templates can be helpful, of course (e.g., van den Akker et al., 2021), but I would argue that the real challenge is conceptual: We need to disentangle the various practices that were conflated in basic experimental preregistration.

To start a broader conversation, I will describe what I believe to be helpful for close relationships researchers who are interested in preregistration. Critically, I start by assuming that we can separate out the various amalgamated “preregistration” practices because these practices serve different goals (da Silva Frost & Ledgerwood, 2020; Ledgerwood, 2018, 2019). In most cases, I have found that it is possible to separately consider an easy goal (transparency), an intermediate goal (control Type-I error), and a hard goal (test a theoretical prediction). I tackle the goals in sequence (i.e., only attempt the hard goal if you were also attempting the easy and intermediate goal).

The easy goal is “be transparent about my data analytic decisions.” This goal is achieved by describing various data processing decisions that will likely come up, and by explaining what statistical tests you will run. As best you can, create the table that will ultimately go in the manuscript, and leave placeholders (e.g., x.xx) for the values; this task will likely prompt you to think about all the different ways you could operationalize each variable before you actually start conducting analyses. One can easily do this kind of preregistration in stages: If during data processing you encounter a decision you hadn’t anticipated, decide what to do, then update the preregistration before proceeding further (an example can be found in this article: Eastwick et al., in press). In my experience, this easy goal is almost always achievable.

The intermediate goal is “control my Type-I error.” This is more challenging than a simple two-condition study. It is challenging in part because you may be carrying out many tests of a hypothesis (e.g., you are examining the effects of 3 correlated independent variables on 2 correlated dependent variables); in such cases, it will be important to specify how you will control family-wise Type-I error rate. I tend to favor a Holm-Bonferroni test (Holm, 1979) when it would be conceptually meaningful that some tests might receive support and others would not (e.g., these five traits had significant effects, and these five traits did not). I tend to favor a binomial test (Bahns et al., 2017) when it is conceptually meaningful to have an “up or down” decision about whether the hypothesis was supported across all tests (e.g., across all operationalizations of a given statistical interaction, the number of significant effects did not exceed what would be expected due to chance). A related challenge is that you might have existing knowledge about a dataset. Here, I suggest transparently reporting what you know about the dataset already; it is not clear whether there is a formal way of controlling Type-I error in this circumstance, but at least readers would be able to evaluate for themselves if your effects are overfitted.

The hard goal is “test an a priori prediction derived from theory.” For many scholars, this is exactly what preregistration means: recording a prediction ahead of time (Nosek et al., 2018). In direct contradiction to this convention, I tell close relationships researchers to avoid the word “predict” entirely when addressing the easy and intermediate goals above; only use that word to address the hard goal (and there is nothing wrong about addressing only the easy and intermediate goals). In the large intensive datasets used by close relationships researchers, prediction can sometimes boil down to a person’s own intuitive hunch. There is nothing wrong with recording these hunches ahead of time—you will learn about how good your hunches are!—but your a priori hunches do not affect Type-I error, nor do they test a theory (da Silva Frost & Ledgerwood, 2020; Ledgerwood, 2018, 2019). If you want to test an a priori prediction, describe how the hypothesis derives from a theory, and describe how you (or another reasonable scholar) would adjust your confidence in the theory upwards or downwards depending on the results. I usually only advocate the pursuit of the hard goal if the study design permits you to adjust your confidence up or down; long-shot, counterintuitive studies that are interesting only if they “work” are usually bad candidates for the hard goal of testing an a priori prediction derived from theory.

We need to pull apart the myriad goals that preregistration can achieve and devise better guidelines for scholars who work with datasets which do not easily conform to the basic experimental preregistration schema. I find that the tripartite structure described above gets 80% of the way there.

5. We would like to see: The field getting serious about transparency through a new approach to pre-registration and a wider range of Registered Report opportunities (from JB). One significant change I would like to see with the aim of increasing transparency is a new approach to pre-registration and the opportunities to submit to peer-reviewed outlets in the form of Registered Report. I think of pre-registration as a contract that scientists have with the community. I am open to reading any analyses that authors have conducted on their data; however, I do want to know at what stage they planned to conduct those analyses and whether they already knew other results at the time. The adoption of Dr. Eastwick’s easy goal described earlier in this article is a great way to encourage pre-registration that is not too restrictive. For example, my lab regularly pre-registers analyses in addition to those focused on testing our hypotheses. In our pre-registration document, we often include analyses we know we want to conduct to explore the data but may not have a specific hypothesis we want to test. Or we may want to conduct follow up tests to more fully understand the data set but these analyses sometimes depend on initial results. Therefore, we label the analyses as exploratory or note that we plan to conduct a series of analyses if the initial results conform to a specific pattern. Then, it is clear what was (and was not) planned as in cases where scientists conduct follow-up studies based on an exploratory finding (e.g., Munin & Beer, 2022) or are asked to do additional analyses (e.g., Freedman et al., 2018).

I also expect that editors and reviewers will take pre-registration seriously rather than ignoring it or asking authors if they really did what they pre-registered unless there is some type of discrepancy in the record. There has to be some degree of trust with pre-registration—otherwise, we run the risk of undermining the use of archival data sets. We cannot argue that people should focus on data sharing on the one hand but then make it difficult for people to publish from archival data on the other hand because they cannot really “prove” that they did not conduct any analyses before posting a public pre-registration analyses plan. Editors and reviewers don’t operate in a vacuum; the others involved in the review need to engage and question why pre-registrations are not honored or ignored to ensure that the field realizes their full value. Without the flexibility to fully delineate a plan or the honoring of the pre-registration in the absence of discrepant information, pre-registrations begin to feel like traps or worthless.

Additionally, I would like to see Registered Reports become the default manuscript submission format at peer-reviewed outlets. It does not need to be the only acceptable format but I would like to see it centered as a primary and widely available format. The restriction to certain types of studies such as only replications and the need to justify a higher bar of novelty or importance than other types of submissions does not align with the improvement of science. If the goal is to evaluate science on the basis of the research question, methods, and analyses rather than whether the results were significant, then we need the primary form of evaluation to match those elements of the research. Additionally, Registered Reports promise to facilitate the dissemination of results rather than hamper them. As many trainees are now seeking careers outside of academia, their later stage research often fails to be published as interest in academic publication can wane quite considerably. However, if trainees were only collecting data with a pending agreement for publication of those results, then results (significant or not) of data collection would be more likely to end up being shared as trainees near graduation would be tasked with adding results and discussion to an accepted manuscript rather than starting the manuscript process from scratch. On a more context-specific note, the adoption of Registered Reports as a standard format would also aid the dissemination of science when access to data collection may be restricted or limited in time. For example, in the early stages of the pandemic, many labs had to freeze data collection. This time could have been spent submitting and reviewing Registered Reports which then could be acted upon as access to data collection become more feasible in between waves of the virus.

6. We would like to see: Enhancing the potential of open data by modeling off of the Love Consortium (from PE). If the goal of making datasets openly available is to allow other researchers to verify original authors’ claims, then the current approach—whereby journals require authors to make a dataset available in a repository for other scholars to inspect—is probably sufficient. But bear in mind that the key innovation of the Love Consortium dataverse is that it was a promotion-focused rather than prevention-focused endeavor. That is, it was formulated to address the question “What more could we be doing with these datasets?” as opposed to “How can we double-check existing publications?” As a result, Dr. Algoe and colleagues designed the Love Consortium dataverse to facilitate collaboration and discovery by bringing different datasets together in a novel and generative way.

Scholars who work with large datasets in other areas of psychology can benefit enormously from this approach. Imagine that a researcher is interested in the correlation between two variables, and they sense from the published literature that this correlation tends to be higher in sample type A than in sample type B. If their subfield maintained a similar dataverse that described all measures administered in each study (whether or not those measures were in the published manuscript), it could be easy for the researcher to search all the datasets that permitted the calculation of this correlation. If enough other datasets contained the correlation of interest, the researcher can calculate more precise effect size estimates than if they had tried to collect the data themselves. In summary, the Love Consortium dataverse illustrates how open data can be leveraged to achieve new discoveries, all the while maximizing statistical power and making efficient use of languishing datasets.

Summary

As three researchers who differ in our career stage, research environment, identities, and subfields, we see many successes of the open science movement but also note that implementation has trailed behind its full potential. We all feel encouraged by the focus on larger samples, greater data sharing, and the use of pre-registration for experimental designs as well as analytical decisions for investigating planned and archival data sets. However, there are a number of areas that will help psychological science move closer to correcting questionable research practices, bolstering replicability, and regaining credibility. Overall, there is potential for the open science movement to become a strong ally in fostering diversity, equity, and inclusion. Additionally, more careful consideration of how to facilitate data sharing and pre-registration could be useful, perhaps accompanied by a shift in the incentive structure of our field.

Funding

Article development was supported by NSF-BCS 2021806 awarded to JSB.

Acknowledgments

The authors have no additional (i.e., non-financial) support to report.

Competing Interests

The authors have declared that no competing interests exist.

References

  • American Psychological Association. (2010). Publication manual of the American Psychological Association (6th ed.). American Psychological Association.

  • Bahns, A. J., Crandall, C. S., Gillath, O., & Preacher, K. J. (2017). Similarity in relationships as niche construction: Choice, stability, and influence within dyads in a free choice environment. Journal of Personality and Social Psychology, 112(2), 329-355. https://doi.org/10.1037/pspp0000088

  • Baumeister, R. F., Vohs, K. D., & Funder, D. C. (2007). Psychology as the science of self-reports and finger movements: Whatever happened to actual behavior? Perspectives on Psychological Science, 2(4), 396-403. https://doi.org/10.1111/j.1745-6916.2007.00051.x

  • Bem, D. J. (2011). Feeling the future: Experimental evidence for anomalous retroactive influences on cognition and affect. Journal of Personality and Social Psychology, 100(3), 407-425. https://doi.org/10.1037/a0021524

  • Beyer, F., Flannery, J., Gau, R., Janssen, L., Schaare, L., Hartmann, H., Nilsonne, G., Martin, S., Khalil, A., Lipp, I., Puhlmann, L., Heinrichs, H., Mohamed, A., Herholz, P., Sicorello, M., & Panagoulas, E. (2021). A fMRI pre-registration template. PsychArchives. https://doi.org/10.23668/PSYCHARCHIVES.5121

  • Botvinik-Nezer, R., Holzmeister, F., Camerer, C. F., Dreber, A., Huber, J., Johannesson, M., Kirchler, M., Iwanir, R., Mumford, J. A., Adcock, R. A., Avesani, P., Backzkowski, B. M., Bajracharya, A., Bakst, L., Ball, S., Barilari, M., Bault, N., Beaton, D., Beitner, J., . . .Schonberg, T. (2020). Variability in the analysis of a single neuroimaging dataset by many teams. Nature, 582, 84-88. https://doi.org/10.1038/s41586-020-2314-9

  • Bowring, A., Maumet, C., & Nichols, T. E. (2019). Exploring the impact of analysis software on task fMRI results. Human Brain Mapping, 40(11), 3362-3384. https://doi.org/10.1002/hbm.24603

  • Bowring, A., Maumet, C., & Nichols, T. E. (2021). Erratum: Exploring the impact of analysis software on task fMRI results. Human Brain Mapping, 42(5), 1564-1578. https://doi.org/10.1002/hbm.25302

  • Buchanan, N. T., Perez, M., Prinstein, M. J., & Thurston, I. B. (2021). Upending racism in psychological science: Strategies to change how science is conducted, reported, reviewed, and disseminated. The American Psychologist, 76(7), 1097-1112. https://doi.org/10.1037/amp0000905

  • Buhrmester, M. D., Talaifar, S., & Gosling, S. D. (2018). An evaluation of Amazon’s Mechanical Turk, its rapid rise, and its effective use. Perspectives on Psychological Science, 13(2), 149-154. https://doi.org/10.1177/1745691617706516

  • Chandler, J. J., & Paolacci, G. (2017). Lie for a dime: When most prescreening responses are honest but most study participants are impostors. Social Psychological & Personality Science, 8(5), 500-508. https://doi.org/10.1177/1948550617698203

  • da Silva Frost, A., & Ledgerwood, A. (2020). Calibrate your confidence in research findings: A tutorial on improving research methods and practices. Journal of Pacific Rim Psychology, 14, Article e14. https://doi.org/10.1017/prp.2020.7

  • Doliński, D. (2018). Is psychology still a science of behaviour? Social Psychological Bulletin, 13(2), Article e25025. https://doi.org/10.5964/spb.v13i2.25025

  • Duker, A., Green, D. J., Onyeador, I. N., & Richeson, J. A. (2022). Managing emotions in the face of discrimination: The differential effects of self-immersion, self-distanced reappraisal, and positive reappraisal. Emotion, 22(7), 1435-1449. https://doi.org/10.1037/emo0001001

  • Eastwick, P. W., Joel, S., Carswell, K. L., Molden, D. C., Finkel, E. J., & Blozis, S. A. (in press). Predicting romantic interest during early relationship development: A preregistered investigation using machine learning. European Journal of Personality. https://osf.io/6zkjt/?view_only=b5e6d8512f604e448acf43749cd874f7

  • Finkel, E. J., Eastwick, P. W., & Reis, H. T. (2015). Best research practices in psychology: Illustrating epistemological and pragmatic considerations with the case of relationship science. Journal of Personality and Social Psychology, 108(2), 275-297. https://doi.org/10.1037/pspi0000007

  • Forscher, P. S., Wagenmakers, E., Coles, N. A., Silan, M. A., Dutra, N. B., Basnight-Brown, D., & IJzerman, H. (2020). The benefits, barriers, and risks of big team science. PsyArXiv. https://doi.org/10.31234/osf.io/2mdxh

  • Freedman, G., Brandler, S., & Beer, J. S. (2018). Does engaging in social rejection heighten or diminish social processing? Comprehensive Results in Social Psychology, 3(3), 224-239. https://doi.org/10.1080/23743603.2019.1684820

  • Giner-Sorolla, R. (2018, January 24). Powering your interaction [Blog post]. Approaching Significance. https://approachingblog.wordpress.com/2018/01/24/powering-your-interaction-2/

  • Goh, J. X., Hall, J. A., & Rosenthal, R. (2016). Mini meta‐analysis of your own studies: Some arguments on why and a primer on how. Social and Personality Psychology Compass, 10(10), 535-549. https://doi.org/10.1111/spc3.12267

  • Goh, J. X., Kort, D. N., Thurston, A. M., Benson, L. R., & Kaiser, C. R. (2019). Does concealing a sexual minority identity prevent exposure to prejudice? Social Psychological & Personality Science, 10(8), 1056-1064. https://doi.org/10.1177/1948550619829065

  • Holm, S. (1979). A simple sequentially rejective multiple test procedure. Scandinavian Journal of Statistics, 6(2), 65-70.

  • Joel, S., Eastwick, P. W., & Finkel, E. J. (2018). Open sharing of data on close relationships and other sensitive social psychological topics: Challenges, tools, and future directions. Advances in Methods and Practices in Psychological Science, 1(1), 86-94. https://doi.org/10.1177/2515245917744281

  • Jones, B. C., DeBruine, L. M., Flake, J. K., Liuzza, M. T., Antfolk, J., Arinze, N. C., Ndukaihe, I. L. G., Bloxsom, N. G., Lewis, S. C., Foroni, F., Willis, M. L., Cubillas, C. P., Vadillo, M. A., Turiegano, E., Gilead, M., Simchon, A., Saribay, S. A., Owsley, N. C., Jang, C., …Coles, N. A. (2021). To which world regions does the valence–dominance model of social perception apply? Nature Human Behaviour, 5(1), 159-169. https://doi.org/10.1038/s41562-020-01007-2

  • Kidwell, M. C., Lazarević, L. B., Baranski, E., Hardwicke, T. E., Piechowski, S., Falkenberg, L.-S., Kennett, C., Slowik, A., Sonnleitner, C., Hess-Holden, C., Errington, T. M., Fielder, S., & Nosek, B. A. (2016). Badges to acknowledge open practices: A simple, low-cost, effective method for increasing transparency. PLoS Biology, 14, Article e1002456. https://doi.org/10.1371/journal.pbio.1002456

  • Klein, R. A., Ratliff, K. A., Vianello, M., Adams, R. B., Jr., Bahník, Š., Bernstein, M. J., Bocian, K., Brandt, M. J., Brooks, B., Brumbaugh, C. C., Cemalcilar, Z., Chandler, J., Cheong, W., Davis, W. E., Devos, T., Devos, T., Eisner, M., Frankowska, N., Furrow, D., . . .Nosek, B. A. (2014). Investigating variation in replicability. Social Psychology, 45(3), 142-152. https://doi.org/10.1027/1864-9335/a000178

  • Lai, C. K., Skinner, A. L., Cooley, E., Murrar, S., Brauer, M., Devos, T., Calanchini, J., Xiao, Y. J., Pedram, C., Marshburn, C. K., Simon, S., Blanchar, J. C., Joy-Gaba, J. A., Conway, J., Redford, L., Klein, R. A., Roussos, G., Schellhaas, F. M. H., Burns, M., . . .Nosek, B. A. (2016). Reducing implicit racial preferences: II. Intervention effectiveness across time. Journal of Experimental Psychology. General, 145(8), 1001-1016. https://doi.org/10.1037/xge0000179

  • Ledgerwood, A. (2017, January 24). Why the f*ck I waste my time worrying about equality [Blog post]. Incurably Nuanced. http://incurablynuanced.blogspot.com/2017/01/inequality-in-science.html

  • Ledgerwood, A. (2018). The preregistration revolution needs to distinguish between predictions and analyses. Proceedings of the National Academy of Sciences of the United States of America, 115(45), E10516-E10517. https://doi.org/10.1073/pnas.1812592115

  • Ledgerwood, A. (2019). New developments in research methods. In E. J. Finkel & R. F. Baumeister (Eds.), Advanced social psychology (2nd ed., pp. 39–61): Oxford University Press.

  • Ledgerwood, A., Hudson, S. T. J., Lewis, N. A., Jr., Maddox, K. B., Pickett, C. L., Remedios, J. D., Cheryan, S., Diekman, A. B., Dutra, N. B., Goh, J. X., Goodwin, S. A., Munakata, Y., Navarro, D. J., Onyeador, I. N., Srivastava, S., & Wilkins, C. L. (2022). The pandemic as a portal: Reimagining psychological science as truly open and inclusive. Perspectives on Psychological Science, 17(4), 937-959. https://doi.org/10.1177/17456916211036654

  • Lucas, R. E., & Donnellan, M. B. (2021). Introduction to the special issue: Preregistered studies of personality development and aging using existing data. The Journals of Gerontology. Series B, Psychological Sciences and Social Sciences, 76(1), 1-5. https://doi.org/10.1093/geronb/gbaa192

  • Matias, J. N., Lewis, N., & Hope, E. (2021, September 7). Universities say they want more diverse faculties. So why is academia still so White? FiveThirtyEight. https://fivethirtyeight.com/features/universities-say-they-want-more-diverse-faculties-so-why-is-academia-still-so-white/

  • Moshontz, H., Campbell, L., Ebersole, C. R., IJzerman, H., Urry, H. L., Forscher, P. S., Grahe, J. E., McCarthy, R. J., Musser, E. D., Antfolk, J., Castille, C. M., Evans, T. R., Fiedler, S., Flake, J. K., Forero, D. A., Janssen, S. M. J., Keene, J. R., Protzko, J., Aczel, B., Solas, S. Á., …Chartier, C. R. (2018). The Psychological Science Accelerator: Advancing psychology through a distributed collaborative network. Advances in Methods and Practices in Psychological Science, 1(4), 501-515. https://doi.org/10.1177/2515245918797607

  • Murphy, M. C., Mejia, A. F., Mejia, J., Yan, X., Cheryan, S., Dasgupta, N., Destin, M., Fryberg, S. A., Garcia, J. A., Haines, E. L., Harackiewicz, J. M., Ledgerwood, A., Moss-Racusin, C. A., Park, L. E., Perry, S. P., Ratliff, K. A., Rattan, A., Sanchez, D. T., Savani, K., …Pestilli, F. (2020). Open science, communal culture, and women’s participation in the movement to improve science. Proceedings of the National Academy of Sciences of the United States of America, 117(39), 24154-24164. https://doi.org/10.1073/pnas.1921320117

  • Munin, S., & Beer, J. S. (2022). Does attention to one’s own emotion relate to the emotional interpretation of other people’s faces? Collabra. Psychology, 8(1), Article 34565. https://doi.org/10.1525/collabra.34565

  • Nelson, L. D., Simmons, J., & Simonsohn, U. (2018). Psychology’s renaissance. Annual Review of Psychology, 69, 511-534. https://doi.org/10.1146/annurev-psych-122216-011836

  • Nosek, B. A., Ebersole, C. R., DeHaven, A. C., & Mellor, D. T. (2018). The preregistration revolution. Proceedings of the National Academy of Sciences of the United States of America, 115(11), 2600-2606. https://doi.org/10.1073/pnas.1708274114

  • Open Science Research Consortium. (2015). Estimating the reproducibility of psychological science. Science, 349(6251), https://doi.org/10.1126/science.aac4716

  • Richeson, J. A. (2020, September). Americans are determined to believe in Black progress: Whether it’s happening or not. The Atlantic. https://www.theatlantic.com/magazine/archive/2020/09/the-mythology-of-racial-progress/614173/

  • Roberts, S. O., Bareket-Shavit, C., Dollins, F. A., Goldie, P. D., & Mortenson, E. (2020). Racial inequality in psychological research: Trends of the past and recommendations for the future. Perspectives on Psychological Science, 15(6), 1295-1309. https://doi.org/10.1177/1745691620927709

  • Rosnow, R. L., & Rosenthal, R. (1989). Statistical procedures and the justification of knowledge in psychological Science. The American Psychologist, 44(10), 1276-1284. https://doi.org/10.1037/0003-066X.44.10.1276

  • Simmons, J. P., Nelson, L. D., & Simonsohn, U. (2011). False-positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychological Science, 22(11), 1359-1366. https://doi.org/10.1177/0956797611417632

  • Simmons, J. P., Nelson, L. D., & Simonsohn, U. (2017, November 6). How to properly preregister a study. Data Colada. http://datacolada.org/64

  • Simonsohn, U. (2013). Just post it: The lesson from two cases of fabricated data detected by statistics alone. Psychological Science, 24(10), 1875-1888. https://doi.org/10.1177/0956797613480366

  • Srivastava, S. (2014, May 25). Does the replication debate have a diversity problem? The Hardest Science. https://thehardestscience.com/2014/05/25/does-the-replication-debate-have-a-diversity-problem/

  • Syed, M., & Kathawalla, U. K. (2022). Cultural psychology, diversity, and representation in open science. In K. C. McLean (Ed.), Cultural methods in psychology: Describing and transforming cultures (pp. 427-454). Oxford University Press. https://doi.org/10.1093/oso/9780190095949.003.0015.

  • Tabibnia, G., Satpute, A. B., & Lieberman, M. D. (2008). The sunny side of fairness: Preference for fairness activates reward circuitry (and disregarding unfairness activates self-control circuitry). Psychological Science, 19(4), 339-347. https://doi.org/10.1111/j.1467-9280.2008.02091.x

  • van den Akker, O., Weston, S., Campbell, L., Chopik, B., Damian, R., Davis-Kean, P., Hall, A., Kosie, J., Kruse, E., Olsen, J., Ritchie, S., Valentine, K. D., Veer, A. V., & Bakker, M. (2021). Preregistration of secondary data analysis: A template and tutorial.  Meta-Psychology5, Article 2625. https://doi.org/10.15626/MP.2020.2625

  • Van Horn, J. D., Grafton, S. T., Rockmore, D., & Gazzaniga, M. S. (2004). Sharing neuroimaging studies of human cognition. Nature Neuroscience, 7(5), 473-481. https://doi.org/10.1038/nn1231

  • Vanpaemel, W., Vermorgen, M., Deriemaecker, L., & Storms, G. (2015). Are we wasting a good crisis? The availability of psychological research data after the storm. Collabra, 1(1), Article 3. https://doi.org/10.1525/collabra.13

  • Wagenmakers, E.-J., Wetzels, R., Borsboom, D., van der Maas, H. L., & Kievit, R. A. (2012). An agenda for purely confirmatory research. Perspectives on Psychological Science, 7(6), 632-638. https://doi.org/10.1177/1745691612463078

  • Whitaker, K., & Guest, O. (2020). #bropenscience is broken science: Kirstie Whitaker and Olivia Guest ask how open ‘open science’really is. The Psychologist, 33, 34-37.

  • Wicherts, J. M., Borsboom, D., Kats, J., & Molenaar, D. (2006). The poor availability of psychological research data for reanalysis. The American Psychologist, 61(7), 726-728. https://doi.org/10.1037/0003-066X.61.7.726