Skip to main content
F1000Research logoLink to F1000Research
. 2019 Nov 25;8:971. Originally published 2019 Jun 26. [Version 2] doi: 10.12688/f1000research.19619.2

Preprints and Scholarly Communication: An Exploratory Qualitative Study of Adoption, Practices, Drivers and Barriers

Andrea Chiarelli1,a, Rob Johnson1, Stephen Pinfield2, Emma Richens1
PMCID: PMC6961415  PMID: 32055396

Version Changes

Revised. Amendments from Version 1

We have now updated our article based on reviewers' feedback. First of all, we have updated our title to clearly reflect that this is a qualitative study.  In a number of sections, we have made changes to better balance our findings or clarify them when it comes to the benefits and challenges in the preprints landscape. This included the addition of new literature sources as recommended by the reviewers, which we thankfully acknowledge. New literature published since our original submission has also been incorporated. Finally, we have enriched our supplementary data and have provided a tabular version of our final recommendations (Figure 1).

Abstract

Background: Since 2013, there has been a dramatic increase in the number of preprint servers. Little is known about the position of researchers, funders, research performing organisations and other stakeholders with respect to this fast-paced landscape. In this article, we explore the perceived benefits and challenges of preprint posting, alongside issues including infrastructure and financial sustainability. We also discuss the definition of a ‘preprint’ in different communities, and the impact this has on uptake.

Methods: This study is based on 38 semi-structured interviews of key stakeholders, based on a purposive heterogeneous sampling approach and undertaken between October 2018 and January 2019. Interviewees were primarily drawn from biology, chemistry and psychology, where use of preprints is growing. Interviews were recorded, transcribed and subjected to thematic analysis to identify trends. Interview questions were designed based on Innovation Diffusion Theory, which was also used to interpret our results.

Results: Participants were conscious of the rising prominence of preprints and cited early and fast dissemination as their most appealing feature. Preprints were also considered to enable broader access to scientific literature and increased opportunities for informal commenting. The main concerns related to the lack of quality assurance and the ‘Ingelfinger rule’. We identified trust as an essential factor in preprint posting, and highlight the enabling role of Twitter in showcasing preprints.

Conclusions: The preprints landscape is evolving fast, and disciplinary communities are at different stages in the innovation diffusion process. The landscape is characterised by experimentation, which leads to the conclusion that a one-size-fits-all approach to preprints is not feasible. Cooperation and active engagement between the stakeholders involved will play an important role going forward. We share questions for the further development of the preprints landscape, with the most important being whether preprint posting will develop as a publisher- or researcher-centric practice.

Keywords: preprints, scholarly communication, peer-review, innovation diffusion theory

Introduction

Background

The period since 2013 has seen a marked rise in the number of preprint servers set up for different communities in order to facilitate the rapid dissemination of pre-refereed research outputs. Tennant et al. (2018) list 18 servers launched between 2013 and 2018, variously set up by disciplinary communities, countries, research funders and publishers. One of the first in this new wave was the discipline-based server, bioRxiv – set up by the Cold Spring Harbor Laboratory in 2013 to cover the life sciences – which has been a focus of discussion and debate ( Abdill & Blekhman, 2019; Luther, 2017; Vale, 2015). However, there are a considerable number of other disciplinary servers, including several set up by the Center for Open Science, such as SocArXiv, engrXiv and PsyArXiv (all of which were launched in 2016), as well as platforms such as ESSOAr, set up by the American Geophysical Union in 2018. At the same time, national servers have been launched, including ChinaXiv (for China), IndiaRxiv (for India) and INA-Rxiv (Indonesia) ( Mallapaty, 2019). Funders of research have also set up platforms that enable the sharing of articles before peer-review, including, in 2016, Wellcome Open Research, for Wellcome-funded researchers. In addition, a number of journal publishers have added the dissemination of preprints to their workflows. The open access (OA) publisher, PeerJ, began offering preprint services in 2013, MDPI in 2016 and Cambridge University Press in 2019. Whilst the first of these has now closed its server, significantly it cites its reason for doing so as the change in the preprints landscape between 2013 and 2019: “the academic community is now well-served with other preprint venue options” ( PeerJ, 2019). A number of journals, primarily in biomedical sciences, have adopted a different model, and now deposit submissions from authors in bioRxiv on behalf of authors (where the author agrees to this). Journals practising this model in bioRxiv include Proceedings of the National Academy of Sciences (PNAS), titles published by PLOS and many published by Frontiers ( bioRxiv, n.d.). The F1000Research publishing platform has promoted a novel publication model involving preprints, in which immediate release of author submissions as preprints is followed by open peer review, with revised versions of a paper (alongside author responses to reviewer comments) published in the journal as they are made.

Of course, preprint servers as a venue of scholarly communication are not new. arXiv, the preprint server for physics, mathematics, computer science and related subjects, was set up as early as 1991 ( Larivière et al., 2014) – it is often regarded as an exemplar preprint server, and even the origin of the open access movement (Gajdacs, 2013). RePEc, housing ‘working papers’ in economics, was launched in 1997. There have also been unsuccessful attempts in the past to develop preprints services, for biology, as early as the 1960s ( Cobb, 2017) and again in the late 1990s ( Ginsparg, 2016), and chemistry, in the early 2000s ( Warr, 2003). Nature Precedings, an early publisher-driven preprints server was launched in 2007 and closed in 2012 (Nature Precedings, n.d.).

The move to set up servers since 2013 signals a new level of interest in preprints and a number of recent studies (e.g. Abdill & Blekhman, 2019; Balaji & Dhanamjaya, 2019; Luther, 2017; Tennant et al., 2018) have provided useful overviews of what Tennant et al. (2018, p. 5) call the “explosion of preprint platforms and services”. Significantly, this new wave of preprints has often come from disciplinary communities not previously associated with adoption of preprints. Biomedical disciplines served by bioRxiv, for example, have traditionally been associated with ‘Gold’ open access (publication in journals) rather than ‘Green’ OA (deposit of copies of papers in archives or repositories), and have typically not favoured dissemination of papers in pre-refereed form ( Martín-Martín et al., 2018; Wang et al., 2018). Usage of preprints in these new areas has varied across disciplines and servers but in some cases has been considerable, as evidenced by deposit rates. bioRxiv contained a total of 64,777 items on 12 November 2019. At that same time, ChemRxiv contained 2,892 items and PsyArXiv, 6,629 (although these two servers also contain items in addition to preprints).

The launch of these new preprint servers has led to discussion and debate, and some have suggested that preprints may become a disruptive force in scholarly communication ( Luther, 2017; Velterop, 2018). Green (2019) has argued for a digital transformation of publishing into a two-step process: articles would first be posted as preprints, and then invited to formal peer review only if they receive sufficient attention. He argues that this would not only represent a cost-effective model for OA and drive out predatory journals, but could also resolve the so-called ‘serials crisis’, under which growth in research budgets (which produce articles) consistently outstrips that of library budgets (which are used to purchase articles).

The case has been made for preprints in a number of disciplines, including biology ( Desjardins-Proulx et al., 2013; Fry et al., 2019; Vale, 2015), medicine ( Lauer et al., 2015) and chemistry ( Carà et al., 2017). Some funders have signalled support for preprints being used in grant applications, including National Institutes of Health (NIH) and Chan Zuckerberg Foundation in the USA, and the Wellcome Trust in the UK.

However, sceptics have questioned the value of preprints and even suggested they may be dangerous – circulating versions of articles before they have been quality controlled by peer review may lead to unnecessary risk, particularly in disciplines like medicine ( Krumholz et al., 2018; Leopold et al., 2019; Sheldon, 2018).

This paper aims to explore the current and potential future role of preprints as a vehicle for scholarly communication by investigating current practices, drivers and barriers to their use. The overall objective of the study was to explore the place of preprints in the research lifecycle from the points of view of key actors, including:

  • research funders;

  • research performing organisations;

  • preprint servers and service providers; and

  • researchers (engaged and unengaged).

The topics in focus included usage of preprints, perceived benefits and challenges, policy positions, motivations and strategies. The research took the form of a set of 38 detailed interviews with representatives from these groups.

The study was funded by, and co-produced with, Knowledge Exchange (a group of national organisations from six European countries supporting research infrastructure), as part of their work on open-access policy and service development. It was, therefore, important that the research should not merely have a descriptive purpose but also a prescriptive one, involving setting out possible directions for future policy and action. The study is the first using empirical qualitative data focusing on the new wave of preprint servers set up since 2013, as such it aims to make a significant contribution to knowledge in this dynamic area.

Literature review

Apart from recent discussion on the growth of preprint services ( Abdill & Blekhman, 2019; Balaji & Dhanamjaya, 2019; Tennant et al., 2018), consideration of preprints in the formal academic literature, as well as in the scientific press and other online venues (such as blogs and other social media commentary), has tended to concentrate on five main areas: firstly, defining preprints; secondly, their perceived benefits and challenges; thirdly, disciplinary differences; fourthly, policy developments; finally, the use and impact of preprints. We discuss these in turn in what follows. There are, however, still a relatively small number of empirical studies focusing on preprints, although the body of evidence is now growing rapidly. Nevertheless, much of the literature is still to be found in editorials and opinion pieces rather than data-driven research.

Defining preprints. Different definitions of preprints in the academic literature typically relate to a number of key components: (1) genre, (2) timing, (3) versioning, (4) accessibility, (5) responsibility and (6) value (see Table 1).

Table 1. Components of the definition of a preprint.

Component1. Genre2. Timing3. Versioning4. Accessibility5. Responsibility6. Value
DescriptionThe type of
output that a
preprint is meant
to be – part of
the scientific
literature
The position
of a preprint in
the knowledge
production process
– prior to formal
publication
Preprints as
different versions
of an output at
different stages
of the outputs’
lifecycle – before
peer review
The availability
of preprints
–preprints are
openly available
online
The individual(s)
responsible for
posting preprints
– normally the
author
The usefulness of
preprints to readers
– preprints are
considered valuable
Exemplifying
quotation
“A preprint is
a complete
scientific manuscript.”

Berg et al. (2016, p. 899)
“A ‘preprint’ is
typically a version
of a research paper
that is shared on
an online platform
prior to, or during,
a formal peer
review process.”

(Tennant et al., 2018)
Preprints are
made available
“before, or
in parallel to,
submitting them
to journals for
traditional peer
review.”

Desjardins-
Proulx et al.’s
(2013, p. 1)
A preprint “can
be viewed
without charge
on the Web.”

(Berg et al., 2016, p. 899).
A preprint “is
uploaded by
the authors to a
public server.”

(Berg et al., 2016, p. 899).
“A preprint is a
research output that
has not completed
a typical publication
pipeline but is
of value to the
community and
deserving of being
easily discovered
and accessed.”

(Bourne et al., 2017)
Other
literature
sources
(Balaji & Dhanamjaya, 2019; Bourne et al., 2017; Crossref, 2016;
Luther, 2017; SHERPA/
RoMEO, n.d.)
(Carà et al., 2017;
Johansson et al., 2018; Luther, 2017; Neylon et al., 2017; Rittman, 2018; Sarabipour et al., 2019)
(Bourne et al., 2017; Crossref, 2016)(Johansson et al., 2018;
Sarabipour et al., 2019)

With regard to (1) genre, Berg et al. (2016, p. 899) state, “a preprint is a complete scientific manuscript”, and Bourne et al. (2017) observe, “typically, a preprint is a research article, editorial, review, etc.”. Whilst the latter widen the scope also to include, “a commentary, a report of negative results, a large data set and its description, and more” (p. 1), most of the discourse on preprints tends to assume they are conventional research papers and therefore follow the academic conventions of that ‘genre’ ( Kelly & Autry, 2013).

With regard to (2) timing, the key point made by most commentators is that a preprint is made available before formal publication, which Carà et al. (2017) describe as “prepublication”.

For (3) versioning, the relationship of a preprint to peer review is central. Desjardins-Proulx et al.’s (2013, p. 1) observation is typical in stating that preprints are made available, “before, or in parallel to, submitting them to journals for traditional peer review”. Suber (2012, p. 102) points out that this is not to “bypass peer review”, but that it applies to “works destined for peer review but not yet peer reviewed”. However, Bourne et al. (2017) controversially extend the definition to include “a paper that has been peer reviewed and…was rejected, but the authors are willing to make the content public”.

Accessibility (4) is crucial in definitions. A preprint is normally defined as being (or assumed to be) openly available online: it “can be viewed without charge on the Web” ( Berg et al., 2016, p. 899). The idea of openness is fundamental to discussions on preprints. The venue for distribution of preprints is often assumed to be a freely-accessible server of some kind, a point highlighted by Berg et al. (2016, p. 899), who include in their definition that a preprint “is uploaded by the authors to a public server”.

The above phrase, “by the authors” here is important and relates to (5) responsibility. Responsibility for distribution of preprints is traditionally assumed to be that of the author, a component of the definition that is often implicit in the verbs used to describe dissemination of preprints, such as, “sharing”, “posting” and “self-archiving”.

The final component of (6) value is summarised by Bourne et al. (2017): “a preprint is a research output that has not completed a typical publication pipeline but is of value to the community and deserving of being easily discovered and accessed”. To include that an output is “of value to the community” and is “deserving” of dissemination as part of the definition of what constitutes a preprint is interesting, since it includes a judgement of value in the definition. It would be difficult to demonstrate the value of each deposit as it is made. The idea of value is, however, one that is implicit in much of the discourse on preprints.

With ambiguities associated with each of these six definitional components, Neylon et al. (2017) are right that “no universal definition of preprints exists”. The label itself is ambiguous, composed as it is of ‘pre’ and ‘print’. The ‘pre’ of ‘preprint’ has sometimes been defined in relation to formal publication, with a preprint characterised as “prepublication”, leading to the controversial question of whether a preprint can itself be considered a ‘publication’ in its own right ( Larivière et al., 2014). More commonly, however, the ‘pre’ part of ‘preprint’ specifically refers to peer-review and is contrasted with ‘postprint’, a version produced after peer review. The conflation of peer review and publication in some discussions is a reflection of their close association in scholarly communication. It is interesting that the use of the terminology of ‘postprint’ in contradistinction to ‘preprint’, and with both termed generically as ‘eprints’, has declined in recent years. However, Tennant et al. (2018) have proposed its revival for reasons of clarity. Of course, the ‘print’ part of ‘preprint’ is largely anachronistic, but like terms such as ‘paper’ and ‘manuscript’, has continued to be used even in a digital environment.

Perceived benefits and challenges of preprints. Of the advantages of preprints discussed in the literature, perhaps the most prominent are the early and rapid dissemination of research results ( Khera, 2018). Using preprints has the potential to “accelerate” science, something that is particularly useful, for example, in combatting outbreaks of diseases ( Johansson et al., 2018). The formal scientific publication process is often seen as frustratingly slow, particularly in a context where final versions of articles may be little different from preprints ( Klein et al., 2016). Preprints allow authors to assert priority early – a preprint is date-stamped in a way widely recognised by many communities ( Ginsparg, 2016; Mallapaty, 2019; Tennant et al., 2019). Preventing researchers being ‘scooped’ is a major priority in many fast-moving disciplines, but applies, at least to some extent, in all areas of academic research, where novelty is prized. Early dissemination is seen by some as especially useful to a number of members of the scholarly community in particular, with early career researchers (ECRs) commonly identified as specific potential beneficiaries, as preprints can allow them to rapidly achieve “visibility” and demonstrate productivity in job and grant applications ( Desjardins-Proulx et al., 2013; Sarabipour et al., 2019; Tennant et al., 2019).

As well as being a fixed point in the scholarly discourse (date-stamped, etc.), another benefit of preprints emphasised in the literature is in the fact they are still subject to change. Authors can benefit from what Pinfield (2004) has called “informal peer review” of versions of their papers. Ginsparg calls this “crowdsourced peer review”, in contrast to “journal-mediated peer review”, and states, “authors benefit greatly from feedback from interested readers, contributing to improved versions of articles, which are then uploaded” ( Ginsparg, 2016, p. 5). There are, however, few empirical studies of such feedback and its value. There also appears to be little acknowledgement that the claim, when used to make the case for preprints in general terms, stands in tension to the one cited earlier that preprints often differ little from final published versions.

Other key advantages of preprints include wider and fairer distribution of research results, both within and beyond the academic community, something fundamental to many arguments for openness in general ( Ginsparg, 2016; Sarabipour et al., 2019). Access to preprints for machine-based crawling in order to facilitate text mining is also seen as an advantage by some commentators ( Chodacki et al., 2017). Furthermore, preprints, partly as a result of wider dissemination, can also increase numbers of citations of papers ( Davis & Fromerth, 2007) and create opportunities for collaborations ( Kleinert et al., 2018). Finally, preprint servers, their advocates argue, can sometimes usefully also house research outputs that might otherwise be ‘homeless’, including items that do not end up being published in peer-reviewed journals ( Bourne et al., 2017).

Perhaps the most prominent criticism of preprints relates to this last issue: the lack of quality assurance through peer review ( Sheldon, 2018). As well as a general concern about lowering quality standards, lack of quality control has been seen as potentially “dangerous” in, for example, areas such as medicine ( Krumholz et al., 2018) as “reports that have not undergone formal peer review [organised by a journal] could be misleading” ( Lauer et al., 2015). Furthermore, uncertified science might be reported prematurely in the media and might even give rise to ‘fake news’ ( Sheldon, 2018). Some insist that, at the very least, the opportunity to disseminate knowledge rapidly without peer review may encourage academics to produce low-quality outputs on fashionable topics ( Teixeira da Silva, 2017). This issue, however, might be mitigated by the fact that authors sharing incorrect or low-quality research are at risk of reputational damage ( McGlynn, 2017), which could affect their career and future prospects.

Despite claims of the value of informal peer review enabled by preprints, some have pointed to the limited use of commenting and feedback features on preprint servers ( Sarabipour et al., 2019). Others have gone further and questioned the value of self-appointed reviewers, as opposed to those selected by journal editors (see the issue of “self-policing” highlighted by Harnad, 1998). Preprint posting, however, is not normally seen as a substitute for peer review, currently managed by journals, in filtering content ( Suber, 2012), a process that is commonly valued, even if recognised to be imperfect ( Lee et al., 2013).

A number of authors report the concern that dissemination of a preprint may be considered ‘prior publication’, thereby jeopardising acceptance of the paper by a journal – the so-called ‘Ingelfinger rule’ ( Nallamothu & Hill, 2017). Whilst this convention has come under criticism and been withdrawn by many publishers, it still exists for some journals, e.g. in medicine and chemistry ( Lauer et al., 2015; Leopold et al., 2019; Teixeira da Silva & Dobránszki, 2019).

It is noticeable that the literature on the pros and cons of preprints has, like many aspects of open science, given rise to robust discussion and debate. The paper by Krumholz et al. (2018) cited above is itself structured as a debate, with the first two authors making the case for preprints and the third expressing concerns. Sheldon’s (2018) opinion piece in Nature arguing that preprints could have a negative impact beyond the scientific community, was met with vociferous rebuttals in the letters pages of the journal the following month ( Fraser & Polka, 2018; Sarabipour, 2018; Tennant et al., 2018). On social media, such as Twitter, there have also been vigorous exchanges (e.g. Twitter, 2019).

Disciplinary differences. Disciplinary differences have been a prominent feature of the literature on preprints. Neylon et al. (2017) in their seminal work conceptualising preprints usefully distinguish between the “state” and “standing” of preprints. “State” relates to, “external, objectively determinable, characteristics” of preprints; “standing” refers to the “position, status, or reputation” of preprints. Neylon et al. (2017) discuss in detail how preprints of similar states can have very different standings in different disciplinary communities, using the example of the contrast between physics and life sciences. For example, it is not universally agreed when an output should be citable (in the literature, funding proposals or promotion cases) or when it can be used to establish a claim of precedence. In disciplines where a preprint is not considered appropriate to establish precedence, it has also been suggested that making a preprint available may actually encourage research to be scooped by rival researchers who publish in a recognised journal before the preprint authors (the ‘flip side’ of the priority claim argument above) ( Kaiser, 2017).

It is commonly observed that physics has a well-established preprint tradition unlike many other Science, Technology and Medicine (STM) disciplines. Lauer et al. (2015, p. 2448) note, “biology…has trailed behind, whereas clinical research remains well behind.” Carà et al. (2017, p. 7924) characterise chemistry as being “late in embracing preprints”. Such language (“behind”, “late”) seems to represent an assumption that all disciplines will eventually come to use preprints, and that different disciplines are now simply at different points in the adoption process. Such a view has been disputed, however, with some arguing disciplinary differences in communication practices are likely to exist in the long term and therefore that preprints will not be adopted universally across disciplines ( Kling & McKim, 2000).

Policy developments. Of course, disciplinary practices do not operate in a vacuum. They are influenced, amongst other things, by the policy environment in which researchers work. Policies affecting researchers’ practices are developed by a number of groups: publishers, funders and institutions. Publisher policies are critical, with the Ingelfinger rule and deposit embargoes being examples of key policies that may have a negative impact on uptake in use of preprints. To this position of publishers resisting use of preprints, may now be added the contrasting recent development of some publishers embracing preprints, even setting up their own preprints services ( Callaway, 2013; Cambridge University Press, 2019). This development is not completely unprecedented, however, since it does build to some extent on well-established processes in areas like high-energy physics of integrating preprints into the journal submission process (as an example, some physics journals allow submission by simply pointing to an arXiv preprint).

Perhaps the most noticeable shift recently in terms of policy is that of funder policies. Some funders have now explicitly signalled support for use of preprints, including allowing citation of preprints in funding bids, and support their inclusion in cases for academic advancement ( Berg et al., 2016; Bourne et al., 2017). Very few funders, however, currently mandate use of preprints, although this has been proposed by preprints advocates ( Sever et al., 2019).

Institutional policy in this area shows some limited movement, with examples of institutions rethinking (usually rather cautiously) their approaches to criteria for career advancement in relation to the shifting scholarly communication environment, with some explicitly allowing submission of preprints ( ASAPbio, n.d.). It appears that, at present, many organisations still rely on metrics such as the journal impact factor when it comes to review, promotion and tenure of their staff ( McKiernan et al., 2019). Some have argued that initiatives such as Declaration on Research Assessment (DORA), with its emphasis on the quality of the output rather than venue of publication, promote use of preprints ( Polka, 2018), including acceptance of preprints are part of institutional researcher evaluation processes. Another interesting area of institutional policy is the positioning of the institutional repository (IR) in relation to preprints. IR policies and practices differ in this area, with many to date having focused on versions of outputs following peer review, although that is not a universal position ( Baughman et al., 2018).

The use and impact of preprints. A noticeable recent development in the literature has been publication of a number of empirical studies on the use and impact of preprints. These include Carneiro et al.’s (2019) study which compared the quality of reporting of findings in preprints from PubMed and bioRxiv against formally-published journal articles based on a number of criteria tested through a questionnaire. They found that the “quality of reporting in preprints in the life sciences is within a similar range as that of peer-reviewed articles, albeit slightly lower on average, supporting the idea that preprints should be considered valid scientific contributions”. Abdill & Blekhman (2019) analysed the growth in submissions to bioRxiv. Their work also shows a positive correlation between the use of papers on bioRxiv (measured by downloads) and the impact factor of the journal in which papers were subsequently published. Most recently, papers by Fraser et al. (2019) and Fu & Hughey (2019) have found evidence of a citation and altmetric score advantage for papers deposited in bioRxiv compared with those not made available as preprints. All of these studies were made available as preprints on bioRxiv in 2019 and, interestingly, all focus on bioRxiv, evidence of current interest in this growing service. They all notably present evidence of the positive impact of preprints. The papers usefully add to the growing empirical evidence base on preprints, augmenting studies such as Klein et al. (2016), on arXiv.

Methods

Analytical framework

As preprints and preprint servers are still innovative developments for most disciplines, it is important to gain an in-depth understanding of the perspectives of different stakeholders. In order to explore issues, such as varying motivations, differing behaviours, and conflicting perspectives, particularly in emerging areas, qualitative research methods are often deployed, since they are well-suited to such investigations. We chose to carry out detailed interviews of key actors in this space who could explain in depth their perceptions, attitudes and practices in relation to preprints. Participants were asked about their perspectives on preprints in general, but we intentionally recruited interviewees (where they had disciplinary affiliations) particularly from disciplines where preprint services are relatively new and rapidly growing. These were biology, chemistry and psychology, corresponding to the preprint servers bioRxiv, ChemRxiv and PsyArXiv. Focusing on these areas helped us to gauge the impact that preprints are having in areas where they are more innovative, and since many of our participants were able to speak more generally about preprints, we were able to draw comparisons with disciplines where preprints are established and which are better represented in the literature (e.g. physics, computer science, and economics).

As a way of framing our research design, we used innovation diffusion theory (IDT), a well-established theoretical framework for explaining the way innovations are adopted in different contexts ( Rogers, 2003). IDT has been tested and deployed widely and proved to be a robust explanatory model in a range of contexts, including OA ( Hampson, 2014; Jones et al., 2006; Pinfield & Middleton, 2016; Xia, 2012). It is designed to describe “the process by which an innovation is communicated through certain channels over time among members of a social system” ( Rogers, 2003) (original emphasis). An innovation is defined as “an idea, practice, or object that is perceived as new by an individual or other unit of adoption” ( Rogers, 2003). Preprints are both cultural innovations, as they aim to change practices in scholarly communication, and technological innovations, in terms of changes to infrastructures and processes. IDT offers ways in which these aspects of preprints as innovation can be understood, particularly in relation to two key issues: the “innovation decision process”, and the “rate of adoption”.

The innovation adoption decision process is seen as going through a number of consecutive steps:

  • 1.

    Knowledge, when the decision-making unit is exposed to the innovation’s existence and gains an understanding of how it functions;

  • 2.

    Persuasion, when the decision-making unit forms a favourable or unfavourable attitude towards the innovation;

  • 3.

    Decision, when the decision-making unit engages in activities that lead to a choice to adopt or reject the innovation;

  • 4.

    Implementation, when the decision-making unit puts an innovation into use; and

  • 5.

    Confirmation, when the decision-making unit seeks reinforcement for an innovation-decision already made but may reverse the decision if exposed to conflicting messages about it. ( Rogers, 2003)

A particularly important concept to understand the success of innovations is their rate of adoption (see Table 2). This was used as the initial basis of the design of the interview questions. From the key factors associated with the rate of adoption, we selected a range of features that appear appropriate for the scope of the present investigation and that were suitable to discuss via interviews for the different stakeholder groups:

Table 2. Variables affecting the rate of adoption of an innovation (adapted from Rogers, 2003).

Variables determining the rate of
adoption
Components of variables
Perceived attributesRelative advantage
Compatibility
Complexity
Trialability
Observability
Nature of the social systemNorms
Degree of network interconnectedness
Extent of change agents’ promotion effortsPromotion efforts
Communication channelsMass media
Interpersonal
Type of innovation-decisionOptional
Collective
Authority
  • Perceived attributes, are what stakeholders feel are the benefits arising from an innovation, in this case the introduction of preprints. Perceived attributes can be split into relative advantage, compatibility, complexity, trialability and observability.

  • Nature of the social system, including “norms”, which are the established behaviour patterns for the members of a social system. They define a range of accepted behaviours and serve as a guide or standard for the behaviour of members of a social system. Norms tell individuals what behaviour they are expected to adopt and may be affected by the introduction of an innovation and by the actions of change agents within the social system. These relate to the level of interconnectedness or cohesiveness of the community.

  • Change agents’ promotion efforts, which are the efforts made by individuals with influence in the system to promote the adoption of an innovation deemed desirable by a change agency (e.g. funders and institutions, service providers, publishers). Change agents often use opinion leaders in a social system as their lieutenants in diffusion activities.

  • Type of innovation decision, which describes how the uptake of preprints is affected when individuals or communities support them, or authorities mandate their posting.

The topic of communication channels (additionally part of IDT theory on the rate of adoption) also arose organically from the discussions with our interviewees.

Interview sampling and approach. Interview questions were developed based on the factors outlined in Table 2 (see Extended data ( Chiarelli et al., 2019a) for more information). From an initial long list of possible questions, areas for investigation were prioritised based on the different stakeholder groups involved, and our review of the literature. We also incorporated questions associated to current policy-related issues, agreed in consultation with the Knowledge Exchange steering group, taking into account the innovation adoption process. This ensured that the approach taken was both theoretically robust and sufficiently grounded in practice to be useful in generating actionable insights. Interviews were conducted using a semi-structured approach – incorporating a ‘spine’ of common questions for all participants, and some questions specific to different actor groups – allowing room for the interviewer to pursue areas of interest arising from participant responses, including probing for greater clarity, where needed (Bryman, 2015).

The study adopted a heterogeneous purposive sampling approach, aiming to include a wide range of perspectives from actors in the area, selected in a “strategic way” in order to address the objectives of the study (Bryman, 2015). The sample was heterogeneous in a number of respects: firstly, it contained representatives of different roles in the scholarly communication system; secondly, it included participants from different countries (and therefore policy environments); thirdly, it comprised interviewees and with different views and levels of experience of using preprints. Participants comprised senior representatives from research funders, research-conducting organisations (universities and research institutes), preprint services, other related service providers (such as infrastructure providers), as well as researchers, both researchers demonstrably engaged with preprints (they had themselves posted a preprint) and non-engaged (there was no evidence of them having posted a preprint). In this study, we acknowledged the important role of academic publishers but chose not to engage them directly, apart from a sample running preprint services. This decision was made as the publishing community is already discussing preprints in a structured way, for example via the Committee on Publication Ethics (COPE).

Those participants with disciplinary associations (researchers and preprint service providers) came from the disciplines identified (biology, chemistry and psychology) but all participants were asked questions about preprints in general as well as their own community’s experiences. Participants were based in eight countries: Denmark, Finland, France, Germany, Netherlands, Switzerland, UK and USA, all apart from the USA and Switzerland being KE member countries. Participants were identified from the literature and from their associations with relevant services or organisations. Snowball sampling was also used as the research progressed and appeals for participation on social media were also shared (particularly to identify non-engaged researchers). Participants gave their informed consent and agreed to be named as participants in any reporting on the understanding that particular views or quotations reported would not be linked to them or their organisation and that the full text of transcripts would be kept confidential. The research approach adopted by the project was given ethical approval by the University of Sheffield. A full list of participants is available in Chiarelli et al. (2019b).

We undertook 38 semi-structured interviews, with participants distributed across the targeted stakeholder groups as illustrated in Table 3. The sample allowed the study to achieve the desired heterogeneity of actors and perspectives. Interviews took place between October 2018 and January 2019, and ranged from 32–75 minutes in length. They were conducted via GoToMeeting, recorded and fully transcribed using the intelligent verbatim method (including minor edits e.g. removing ‘fillers’ etc). Two interviews took the form of an email Q&A because of restrictions around the participants’ availability. The transcripts were then subjected to thematic analysis ( Braun & Clarke, 2006) which took place in several stages. Initially, members of the research team independently read a sample of transcripts from different stakeholder interviews, including some in common, and then discussed key topics arising from the transcripts. This formed the basis of the initial coding approach then undertaken by E.R. This was reviewed as analysis proceeded, with coding being checked and validated by A.C. and S.P. as it progressed and amended as necessary in light of their comments. The codes were then grouped into themes agreed by the team. These themes form the basis of the findings reported below.

Table 3. Participants by role and country (n=38).

Participants are listed by Chiarelli et al. (2019b).

Stakeholder groupUKGermanyFranceNetherlandsDenmarkUSAFinlandSwitzerlandTotals
Research funders11111-1-6
Research performing organisations12211-1-8
Preprint server providers-1---2-14
Other service providers21---1--4
Engaged researchers32111---8
Unengaged researchers4211----8
Totals11954332138

Limitations and constraints. Like many kinds of qualitative research, this study was designed to be exploratory; in this case, to map out key aspects of the preprints space and suggest policy responses. Our conclusions are tentative. Many of our interviewees were selected because of their knowledge of the issues under investigation, and although our findings based on their views may be transferable to other contexts, they cannot be generalised without further testing, as with most qualitative research. There was a bias in our sampling towards participants aware of and engaged with preprints. Further research, using other methods, will be needed in future in order to generalise across communities as a whole, including non-engaged researchers. Furthermore, some stakeholder groups, such as publishers (who only have very limited representation in this study), and other groups (such as non-academic users of the research literature) could usefully be included in future studies. Our coding was undertaken using agreed protocols and involved a process of validation provided by three different members of the authorial team, but necessarily involves interpretation and judgement on the part of the researchers.

Results

Overview

The analysis of the data identified nine major themes arising from the interviews, which can be grouped into four thematic areas ( Table 4). These themes are used as the framework for presenting results and explored in more detail in what follows.

Table 4. Themes from the interviews.

Thematic areaTheme
Definitions and rolesDefinitions of preprints
Disciplines, cultures and practices
Preprints’ position in the landscape
Potential benefits and challengesPreprints as an asset (benefits)
Preprints as a liability (challenges)
Infrastructure and sustainabilityInfrastructure
Policy
Financial stability and business models
FutureFuture of preprints

Definitions and roles of preprints

Definitions of preprints. In view of the ambiguities and disagreements in the literature around definitions of ‘preprints’, one key aim of the interviews was to ask our participants about their understanding of the term. Whilst all of the participants agreed with our broad definition that ‘a preprint is a research output made available in a form before it has been peer-reviewed and published’, there was considerable variation in the specifics of what that means. Some participants were themselves aware that the term was being used in different ways by different people and there were discrepancies (and in some cases confusion) about its precise meaning. One expressed dislike for the term, saying that it “presupposes…that you are in a print era”. However, other alternatives used by participants in this space, such as “manuscript” and “paper”, as has been observed, are equally anachronistic. There was evidence of participants struggling to find a clear language for the innovation being discussed.

Many saw a preprint as being in a form that was ready to be submitted to a journal, “at the point of submission” (Research performing organisation), as one participant put it. Referring to their own experience, one researcher stated:

  • “the preprint…was only submitted or uploaded at a stage where it was essentially submission-ready for a journal.” (Engaged researcher)

Other participants saw preprints as earlier versions of outputs made available in order to receive comments (e.g. working papers in economics). One participant acknowledged that a preprint was commonly thought of as the submitted version of an article but also discussed possible earlier versions being shared:

  • “…in terms of quality, almost like [the] thing that will appear in the journal later on but if you consider a preprint like a working paper, you…definitely can see it as a step earlier in the whole research process in which there is still the possibility to enrich and improve the later formal publication on the same time.” (Research funder)

Another participant referred to the benefit of this: “it gets feedback from the community” before “official peer review” (Unengaged researcher). A key point made by one participant, but implicit in the comments of many, was that that a preprint is a version of an output that has not been peer reviewed but that the author is “committing to get it peer reviewed” (Other service provider).

  • “…the term itself includes the idea that you’re building it just towards something. That it’s only the preprint and then something will come later after from it.” (Research funder)

Another author said of their own approach to posting preprints: “I always had the intent to publish this in a…proper journal subsequently” (Engaged researcher). A preprint is part of a “continuum” of different research stages, one participant argued, and authors should deposit all versions of their papers (and data) in a repository as part of contributing to that continuum. However, one participant did question this emphasis on a work flow with the preprint being provisional, since it appeared to devalue preprints:

  • “…preprint means there’s something you know, there’s…a paradise afterwards, there’s a better life, and that it’s not a publication of its own.” (Other service provider)

Either as a provisional version of a forthcoming publication or as a “publication” in its own right, a preprint was seen as part of a recognisable genre of scientific output, and which was planned to be part of the formal published literature, but was made available earlier than formal publication.

  • “they’re not that radical, the concept is radical, but when you look at them, they look like articles.” (Other service provider)

Several participants included post-refereed versions of articles within the definition (although some acknowledged these were not “pure” preprints), and others recognised that in reality many authors deposited post-refereed versions on preprint servers, usually to enhance the outputs’ accessibility. A small number of interviewees acknowledged that some papers posted as preprints might not end up being formally published, although this might raise questions about the value of the output:

  • “If there’s nothing to follow the preprint then I would start to wonder what did happen. Why was the work dropped and left on this preprint level?” (Research funder)

Some questioned whether a paper which did not end up being published could legitimately be considered a preprint, with one interviewee asserting, “it’s not a preprint; it’s just a manuscript” (Other service provider). That a preprint is basically a research paper was the assumption of the majority of participants, but one questioned even this:

  • “You know, it’s basically anything, you know between a tweet…or a poster presentation and an actual paper…” (Engaged researcher)

Many acknowledged there was uncertainty about preprints and interviewees were cautious in committing to a definition. In some cases, this related to disciplinary differences – something acknowledged by several interviewees. One interviewee discussed definitional differences between his own discipline, chemistry, and that of physics, mainly in terms of community acceptance. Another participant from the humanities stated:

  • “I don’t think there’s any one perspective that scholars in the humanities have on preprints and I think that there’s some confusion about the terminology.” (Other service provider)

Disciplines, cultures and practices. Disciplinary differences were evident not only in perceptions of what preprints are but also in terms of their acceptance. This was important throughout the interviews. There were firstly differing levels of awareness and, following this, adoption. Some interviewees recognized physics, mathematics and computer science to have well-established preprints practices with very high levels of awareness and adoption, but in other disciplines awareness was often still low:

  • “I think in chemistry it’s small but growing and biology is being helped a lot by bioRxiv …there’s certainly some areas where there is still a kind of much less awareness. I would say really outside the math and physics [communities] the awareness is much lower of what preprints are about.” (Preprint server provider)

One service provider summarized confusions (about definitions and processes) even applying to people trying to submit their work as preprints:

  • “So there’s definitely a growing awareness but it’s still a minority. And we still find that there are some who are confused by the process and when they submit a preprint they don’t really understand. Despite everything we try and make them aware of, they don’t really understand the process.” (Preprint server provider)

Furthermore, where there was evidence that awareness was rising, this did not necessarily result in uptake. Some unengaged researchers interviewed were aware what preprints were but had not been motivated to use them to date. For unengaged researchers, there were perceived practical barriers:

  • “I am not entirely sure of the process, that’s the reason why I haven’t done it. I’m sure I could work it out but I’m not entirely sure which preprint server I would use, whether one would be better for my type of work than another…” (Unengaged researcher)

However, there were signs this was beginning to change. One chemist described the situation in their discipline as moving from a position where there had previously been no use of preprints to where use was beginning to happen:

  • “Almost all that is changing now and also from the chemistry part, which might be the related field, there was as far as I know ChemRxiv, which is like the main repository for chemistry data. It has been going on for one or two years maybe.” (Unengaged researcher)

One participant described the process by which awareness of this new development diffuses through the community via informal channels and how they had become more personally aware of it during the course of doing a PhD:

  • “I didn’t get any information on anything like, you know, I found out about it myself, you know, something as simple as that. Most of it’s just through word of mouth, like as you go through a PhD, as you’re talking to people, a lot of meetings, as you hear these terms come up more and more… It’s just accidental.” (Unengaged researcher)

One participant representing a preprint server saw growing awareness and willingness to experiment in different subject communities:

  • “And the momentum behind [name of the preprint server] – and I would not call it a success yet, I would call it momentum – but I think that momentum has given encouragement to other groups of scholars to investigate the possibility of developing a preprint platform for their own discipline. Whether that’s in earth sciences or anthropology or psychology, sociology and so on.” (Preprint server provider)

This was confirmed by another service provider, who stated that before 2016, preprints “were not used much” in their discipline and many researchers “were unaware of what a preprint was”. But this was changing: “the popularity of preprints within the field…is rising” and this was partly attributable to the preprint server that the provider represented which had been a “driving force” for change (Preprint server provider).

The subject specificity of servers was seen as a natural way for preprints services to develop as it was seen as in line with the way researchers worked. However, even within disciplines, some referred to what they saw as significant differences between sub-disciplines, with some engaging in preprints and others not. Participants were agreed, however, that in disciplines where preprints were not established, although there might be some willingness to experiment, there was still a great deal of resistance:

  • “I think lots of my colleagues use them just as much as I do but then I think there are some colleagues that would never post to a preprint.” (Engaged researcher)

Finally, we note that adjacency between disciplines may play an important role. For example, disciplines close to those which posted preprints (e.g. those close to some areas of computer science or physics, which use arXiv consistently) may be more favourably disposed to the practice compared to those from other disciplines.

Preprints’ position in the landscape. Perceptions of the role of preprints in the scholarly communication landscape were partly derived from understandings of and sympathy for open science and open access developments. It was clear that those who were supporters of wider open science developments generally supported increased use of preprints.

We note that peer-reviewed journals are still seen as essential. As such, preprints were for most interviewees “part of a new ecosystem” (Research performing organisation), but not a radical departure from or replacement for selective journals, although their potential to prompt more fundamental change was highlighted by some (see below). The level of integration of preprints in processes associated with submission to a journal was generally seen as low, but there was some awareness of provisions, for example, where a preprint could be transferred to a journal’s submission system. For most disciplines, these were not seen as fundamentally important, however, in determining usage or take up decisions.

One researcher referred to ongoing “scepticism” in “many fields” (Engaged researcher). One of the key reasons for this was that the ‘standing’ of preprints in different fields was seen to be very different. This was reflected, for example, in different perceptions of the value of citing preprints. In some cases, researchers believed that citing preprints was not acceptable in either papers or grant proposals.

Potential benefits and challenges of preprints

Benefits. Participants highlighted a number of (potential) benefits and challenges of preprints which were seen to relate to particular practices around adoption or non-adoption of preprints. For the most part, these correspond to those identified in the literature, but it is useful to see how these are articulated and prioritized by our participants. Table 5 summarises the main benefits of preprints as highlighted in the interviews, comparing them with the literature. We have attempted to rank these by the prevalence of certain points across the entire dataset of interviews and have also classified them as to whether they create benefits at an individual or systemic level, that is whether they benefit individual researchers who practice them or the scholarly communication system as a whole. They are discussed in turn below.

Table 5. Potential benefits of preprints.

Mentions across the entire dataset: ✔*** =over 20 mentions; ✔** =between 10 and 20 mentions; ✔* =fewer than 10 mentions. “Systemic” significance relates to those factors with system-wide impact e.g. the broad scholarly communication system or disciplinary community; “individual” relates to those factors primarily affecting individuals or small groups.

BenefitFocus/ significanceInterviewsLiterature
Early and rapid disseminationSystemic✔***
Increased opportunities for feedbackIndividual ✔***
Preprint servers as an outlet for ‘homeless’ resultsSystemic ✔**
Advantages for early career researchersIndividual ✔**
Preventing scoopingIndividual ✔**
Broader access to scientific researchSystemic ✔**
Increased citation countsIndividual ✔*
Preprints can support collaborationsSystemic ✔*
Preprints in some formats (e.g. xml) and with open
licences are easier to text and data mine
Systemic✔*
Much shorter time before research can be shared,
so authors remain enthusiastic about it
Individual
Preprints may reduce predatory publishingSystemic

Of the benefits highlighted by participants, early and rapid dissemination of research was the most frequent. As one engaged researcher succinctly put it:

  • “The primary purpose of preprints is to communicate scientific knowledge as early as possible to as wide an audience as possible.” (Engaged researcher)

Preprints were described by one researcher as enabling “science in real time” (Engaged researcher), particularly with reference to the lengthy period of peer review and publication that often applied to article publishing. One participant representing a preprint server described the peer review process as, “often long and tortuous” (Preprint server provider), and a university representative described it as a kind of “limbo” (Research performing organisation) in which the research was not being read or used.

Achieving rapid and wide dissemination could in turn “accelerate” the pace of research itself (Other service provider). This might mean being able to “see a result that somebody else did and I can start working on it” (Research funder), even though this type of behaviour was described as appropriate only in very fast-moving sub-disciplines. Research could also make progress thanks to the reduction of “redundant work”, which makes it more likely for researchers to identify “the next big research question” (Research performing organisation) more quickly.

The benefits discussed were mostly seen in terms of communicating with a particular disciplinary community, but some interviewees also mentioned reaching a wider audience, including policymakers or clinicians, in a timely way – research reflecting the latest thinking in an area has more “news value” (Research performing organisation). One engaged researcher also emphasized the benefit of making the latest research available in the case of an outbreak of disease.

Participants were conscious of the importance of dialogue and interchange as part of the research process and so valued the potential for preprints to create opportunity for feedback. Some participants conceived of this from the point of the researcher receiving comments on their paper in order to make corrections – a kind of “debugging” (Engaged researcher) – or other improvements:

  • “…feedback from others to help you with your thinking and to improve your ideas.” (Research performing organisation)

It was commented that some preprint servers facilitate this in various ways, by for example allowing authors to solicit feedback or providing commenting or annotation features. Many respondents saw this in wider systemic terms, rather than just personal: enabling community engagement and discussion – what one funder called “community-oriented discourse on research results” (Research funder). The language of ‘community’ was particularly strong amongst many participants here. Some provided stories of their experience of this, with one researcher commenting on a particular paper that,

  • “the feedback from the community which we received through the preprint was at least as constructive and helpful as the official reviews from the journal.” (Engaged researcher)

Many of the participants agreed that preprints servers could be a useful outlet for otherwise ‘homeless’ research outputs, even though this goes counter to the emphasis of many of preprints as early versions of outputs later formally published elsewhere. The most common sorts of outputs mentioned were null results and replication studies, which would not satisfy journal requirements of novelty or significance. However, participants also mentioned older or under-developed papers that had not been formally published but could easily be deposited on a preprint server and would be of value.

Such an approach could be particularly useful for early career researchers, who benefit in general from posting preprints in order achieve greater “visibility” relatively quickly. This could be particularly useful in the case of funding proposals or job applications in order to demonstrate productivity, although several participants were quick to emphasise that formal publications would be preferable. For all researchers, in fact, preprints were seen as possible evidence of productivity but no real substitute for formal publications.

The benefit of preventing scooping was also prominent – preprints were a way to establish priority:

  • “If you’re an author the benefit of having a preprint in the public domain is it identifies the ideas as being yours, so it prevents you from being scooped because the ideas are down in time stamp against your name.” (Research performing organisation)

However, a small number of participants observed that this might cut both ways, since in disciplines where preprints did not have the standing of a citable resource, making research available in this way might cause the researcher to be scooped: “if you put something online and it’s not yet published [I worry] that it would be stolen by competitors” (Unengaged researcher). The strength of the language (“stolen by competitors”) is telling of the intensely competitive culture in which many researchers work and is in marked contrast to the language of community noted above. However, other researchers were sceptical about whether scooping as a result of sharing preprints was likely to happen.

This was partly because a preprint is available so widely. Breadth of availability was certainly a benefit seen by many participants, primarily in reaching as many members of their own research community as possible. However, many participants also mentioned access by a wider readership, including the general public. Either way, the benefit of increased usage of papers was seen as particularly important. However, once again, the benefits were qualified, with greater value being placed on formal publications or accepted author manuscripts:

  • “I think preprints are a very good kind of second option if you can’t access the published version and you can’t find the author’s accepted manuscript.” (Research performing organisation)

A number of participants were confident making work available as a preprint would increase citations, with the paper being citable in preprint form if available on a preprint server. This applied particularly in fast-moving areas. Getting work out into the community at an early stage also, it was suggested, increased the opportunity to form new collaborations. This is in many respects the more optimistic side of the idea of competitors stealing ideas – the possibility that researchers may be encouraged to develop a collaboration as a result of seeing the developing work of peers.

Challenges. Participants also saw problems with preprints – summarised and compared with the literature in Table 6, with an indication of prevalence, and discussed in turn below. It is worth noting, however, that participants often had a nuanced view of the challenges, commonly stating potential disadvantages but then themselves qualifying their criticisms or citing possible solutions.

Table 6. Potential challenges of preprints.

Mentions across the entire dataset: ✔*** =over 20 mentions; ✔** =between 10 and 20 mentions; ✔* =fewer than 10 mentions. “Systemic” significance relates to those factors with system-wide impact e.g. the broad scholarly communication system or disciplinary community; “individual” relates to those factors primarily affecting individuals or small groups.

ChallengeFocus/SignificanceInterviewsLiterature
Lack of quality assuranceSystemic✔***
Limited use of commenting/feedback features on the serversBoth✔***
Risk of the media reporting incorrect researchSystemic✔***
Possible harm in the case of sensitive areasSystemic✔***
Questionable value of self-appointed reviewersBoth✔***
Information overloadSystemic✔**
The Ingelfinger rule – journals rejecting submissions if they
have been posted as preprints
Individual✔**
Possible reputational damage to the depositor if the preprint
is not good enough
Individual✔*
Possible ‘preprints wars’ in which the findings in one preprint
are quickly attacked in another
Systemic
There may be a rush to post low-quality research about
popular topics
Systemic

Of the problems, the lack of quality of assurance was most widely discussed by participants, with a set of related issues clustering around this theme. In some cases, there was a concern preprints could simply mean lower quality:

  • “…so my worry would be with rapid publication by preprint that there would be an increase in the amount of poor quality science that’s available…” (Unengaged researcher)

However, more commonly, there was the view that a lack of quality assurance meant greater uncertainty and that readers had a greater responsibility to approach preprints critically. Words such as “caution” and “sceptical” were often used. The filtering role of selective journals was valued by many participants. Peer review, for all of its faults (and participants were not slow to point out its possible failings), was still highly valued:

  • “It’s really important that good reviewers have looked at the article and at the results section.” (Research funder)

Peer review was thought of not just as a safety net but as a process which often improved the quality of a paper. Most of the supporters of preprints, did not therefore see them as an alternative to or replacement for peer-reviewed papers but as a complement to them. Whether or not they valued informal comments, participants commonly pointed out that the use of feedback functionality of preprint servers was still limited. One preprint service provider observed that only about 10% of preprints received comments.

Particular concerns were expressed around quality related to the media or members of the public latching onto unreliable findings, and on the harm a preprint containing errors could create in sensitive areas, especially associated with medicine or perhaps law. Several participants observed that people from non-academic contexts may not know the difference between a preprint and peer-reviewed article, and could therefore be more easily misled. Science journalists might potentially play a negative role. One participant said it was common for journalists to report findings from a peer-reviewed paper but “misinterpreting the results…and spreading the news without actually…understanding” the research (Engaged researcher). The risks associated with this were perceived to increase with preprints. Other participants acknowledged the responsibility of journalists, but were more optimistic, citing evidence of journalists using preprint servers responsibly and adding appropriate qualifiers to reports on preprints.

Several participants commented researchers themselves need to be aware of these problems and be cautious about how they published on controversial issues. Peer review did not mean that research papers were immune from such problems in any case.

There was a view expressed that basic screening provided by preprint servers, which was seen as very important by many participants, could address some of these concerns, but was still limited. There was consciousness that such basic checks themselves differed across different preprint servers and also that they would need strengthening in the case of servers dealing with medical outputs. One service provider working in this area suggested that preprint servers in the medical field might have to consider, among other things,

  • “conflict of interest, financial disclosures, assurance that the work reported has been cleared by appropriate ethical review boards and assurance that data underpinning the article is available in an appropriate repository.” (Preprint server provider)

This is a very interesting development which would involve preprint servers undertaking additional quality checks than just the current basic screening, potentially blurring the boundaries between them and peer-reviewed journals.

Participants were conscious of preprints creating what might be called a trust barrier. Whilst some of the determinants of trust for peer-reviewed papers might transfer to preprints (such as the overall shape of the paper, its authors, etc), some of the key determinants (the brand of the journal and its associated peer review processes) were missing. Participants were clear that preprints should be clearly marked as “not peer reviewed” and therefore treated with caution and handled responsibly. However, interviews also indicated possible contributors to trust that might apply to preprints, including: the preprint being widely discussed on social media, receiving comments online (e.g. on the preprint server), being cited, reported on by a magazine or newspaper, recommended by a colleague, and housed in a recognised preprint server.

Whilst there was enthusiasm amongst many participants for “community” based review and commentary on papers facilitated by preprint servers, there was some scepticism about the value of reviewers who “self-select themselves” (Unengaged researcher). People commenting may not understand the area or may use commenting to pursue personal agendas, it was suggested. In addition, an author might invite comments from people who could be expected to be positive about the work. The view was also expressed by some participants that the practice of commenting on preprints could cause difficulty with the formal peer review process, since people who had made public comments may be barred from undertaking blind peer review because of a conflict of interest.

There was concern also about information overload expressed by some participants, some of whom believed that preprints may inflate the number of papers being made available. However, other participants expressed scepticism of this, emphasising that the same number of papers were simply being made available earlier. There was also an acknowledgement that the filtering role played by selective journals was being removed from the process: some participants suggested that this could be at least partially solved with technology-based solutions, improving discoverability and filtering of content:

  • “I think there is a lot of information out there but I think there’s also the potential to find technical solutions that will avoid the information overload.” (Preprint server provider)

A number of participants pointed out that the solutions were better enabled by open, interoperable content, avoiding what one researcher called, “technical and legal restrictions put on by the publishers” (Engaged researcher).

Of these restrictions, the Ingelfinger rule was still the most commonly mentioned (although not by that name) by participants. However, there was considerable uncertainty and confusion amongst some participants (particularly researchers) about what authors could or could not do with regard to depositing preprints, and where they might find reliable information on what was permissible. Several researchers voiced such doubts:

  • “I think there’s always the concern that people are worried that if I put something up there then it restricts where they can submit their paper.” (Engaged researcher)

    “I have a duty to make sure that the work is peer reviewed in the best journals that we can get it into and if I rule certain journals out because I’ve submitted it as a preprint then I’ve kind of done a disservice to myself and those who I work with.” (Unengaged researcher)

Such “fears” themselves acted as a considerable barrier to uptake. Interestingly, one engaged and one unengaged researcher did show an awareness of SHERPA RoMEO (a database of copyright and OA self-archiving policies of academic journals) as a source of information, but this was not common.

Apart from permissions barriers, there were also fears that reputational damage could arise from premature release of preprints. The consequence of this was that sharing a preprint would be delayed until the authors were confident in it to avoid the possibility of any reputational damage. One researcher commented:

  • “I don’t think people in my field would just post off stuff that’s…terrible…because you’re still being judged on what’s going up there.” (Engaged researcher)

In addition, where work was produced by a team of co-authors, gaining agreement from the team was itself an important quality barrier to overcome before disseminating the research in preprint form.

We note that some of the challenges arising from the posting of preprints were only perceived to be substantial in cases where researchers or re-users behave unprofessionally or unethically. However, this does not just apply to preprints, but also to other areas of scholarship and publishing. A key feature of many of the interviews was that participants were rarely able to cite empirical evidence of either benefits or challenges of preprints. At best, personal experience or anecdotes about the experience of colleagues was being cited. As one engaged researcher said:

  • “I don’t have a lot of examples here, but certainly, you know, I hear anecdotes.” (Engaged researcher)

Another said, “I have heard stories on Twitter” (Engaged researcher). It was apparent from many of the interviews that there is a need for further empirical work in this area in order to provide an evidence base for policy and practice.

Infrastructure and sustainability

Infrastructure. The view was commonly expressed by participants who commented on infrastructure that many technologies were already available to support use of preprints. Those mentioned by participants include repository and publishing solutions (open source, such as OSF Preprints, Eprints or DSpace and proprietary technology, such as Figshare), but also the broader scholarly communication infrastructure (e.g. indexing via Crossref). However, several infrastructural issues were emphasised as being important by participants, including technology considerations (such as interoperability, search and discovery), as well as process considerations (including licensing, versioning management, and digital preservation). Functionality and usability considerations for individual services were also mentioned, such as search and annotation of preprints.

Interoperability was regarded as a priority by many participants, with standards often being seen as key. Use of digital object identifiers (DOIs) for preprints (which could then be linked to later versions of the paper) was seen as a particular priority, but other standards such as ORCID, and service providers, such as Crossref, were also mentioned as being important.

Use of standards was seen as an important enabler of effective search and discovery of preprints. Discovery was seen as being achievable largely through network-level discovery services, such as Google Scholar, but interestingly, even greater emphasis was placed on social media, particularly Twitter. It was common for researchers to report finding out about preprints of interest from Twitter, either by following particular individuals or signing up to Twitter feeds set up and managed by preprint servers (including automated Twitter posting when preprints are released). Twitter was reported to be the main way several participants found out about preprints in their field. Several researchers also mentioned gaining feedback on their own preprints as a result of posting links to them on Twitter, with comments sometimes being posted on Twitter itself or other social media rather than the preprint server. The importance of Twitter was also emphasised by service providers, with one representative of a preprint server stating:

  • “I would say that the momentum behind [name of the preprint server] owes a great deal to Twitter, and to Facebook, a bit less so. But nevertheless the effect is there. These are means of amplifying work once it has been posted.” (Preprint server provider)

This is a significant finding. The fact that part of the infrastructure upon which preprint services are currently reliant includes generic discovery services, such as Google Scholar, and social media, particularly Twitter, needs to be taken into account in considering future developments of preprint services. Usage (or at least availability) of Twitter was assumed by participants to be widespread – a reasonable assumption in most Western countries, but not in others, particularly China.

Open licensing was also seen as an important infrastructural issue, which enabled interoperability and discovery. Preprint servers typically offer depositors Creative Commons (CC) licence options and there was some discussion in our interviews of the best one among these. Some authors were clearly confused about the options available to them when depositing their work. Some were aware of various requirements (including from funders) but were also wary of the possible consequences of signing particular licenses on a preprint when it comes to formal publication at a later stage.

Management of versioning was mentioned by a number of researchers although experience of this was mixed, with many authors not taking advantage of the facilities offered by preprint servers in terms of tracking versions. Preprint service providers were divided about whether these services allowed for withdrawal of items, but this was regarded as important by a number of participants in the case of misleading results or disputes between co-authors, for example.

Digital preservation, the final infrastructural issue discussed by participants at any length, was regarded as important in principle but often de-prioritised in practice because of its costly nature.

Policy. Participants identified key policy issues at different levels: publisher, funder and institutional. A cross-cutting issue applying to all policy levels discussed by many participants was, however, the value of preprints as reflected in and recognised by policy. Preprints were recognised to be valuable in providing early open access to research but their value in the scholarly record was commonly qualified by participants. Put simply, they were not considered as valuable as the author accepted manuscript (AAM) or the version of record (VoR), a view reinforced in the minds of many by the fact that many funder and government OA mandates did not have any requirements relating to preprints, but focused rather on AAMs and VoRs. This affected the extent to which different actors regarded preprints as a policy or operational priority. The concept of “standing” ( Neylon et al., 2017) is relevant here, with, as an example, standing relating to whether preprints are seen as an appropriate object for evaluation in exercises such as the UK’s Research Excellence Framework, about which there had apparently been uncertainty. One representative of a university stated:

  • “they’re not acceptable for REF, so they’re not even part of the equation. So it’s the author’s accepted manuscript is the currency we deal with.” (Research performing organisation)

The REF guidance was updated in January 2019 (whilst our interviews were still ongoing) and now states that preprints can be included in REF submissions, but expresses a preference for “final versions” of papers rather than preprints ( Research England, 2019). It is, however, also possible that institutions have developed local conventions for REF submission which in fact discourage submission of preprints, as can be inferred from this quotation.

The value of preprints was seen by many participants not so much in terms of their potential contribution to research evaluation exercises but rather in the extent to which they are a component of a more open research system. Whilst they were seen by some as a useful counterbalance to expensive OA publishing funded by APCs (article processing charges), mostly they were seen as important in terms of their contribution to the overall open science agenda:

  • “…in terms of preprints, I’m more interested in a different problem, which is the problem of the opening up the whole research endeavour throughout the whole research process through the collection of data to the analysis of data to the curation of data through the writing research protocols. And then analysing the…the results and writing the software and all that stuff. Each of those things themselves should be considered a research output. The preprint is towards the end of that.” (Research performing organisation)

Institutional-level policy was seen by many participants in this light. Preprints might be encouraged in a general way by institutional policy, but deposit of preprints would not normally be mandated. In fact, some participants reported that institutional policy was silent on the matter, and in practice preprints were not accepted as deposits to the IR.

Funders, similarly, often encouraged use of preprints and allowed them to be cited in grant applications, but they were not included as an acceptable form of an output which should be made available OA as part of a funder OA mandate. One preprint provider suggested funders might in future mandate use of preprints but pointed to only one small US funder who was currently doing so. There were other possible funders policies that could encourage change:

  • “The other thing that funders really could do is to make very public the fact that they will allow a grantee to cite preprints in her progress reports and then her grant renewal application.” (Preprint server provider)

Publishers’ policies were criticized by many participants as preventing authors from depositing preprints either before submission (the Ingelfinger rule) or after acceptance (contractual exclusions or embargoes). The environment was often seen as a confusing one for authors, something which was itself discouraging.

Financial sustainability and business models. Financial sustainability was a concern for many of the participants with one describing the financial sustainability of a number of preprint services as “fragile” (Research performing organisation), often being based on short-term project funding. One preprint service provider described the work they did in partnership with the Center for Open Science as based largely on the goodwill of “volunteers”. One participant commented that independent preprint servers have no business model associated with them other than grant funding. Although this was sometimes seen as a mark of immaturity, with experimentation in funding and sustainability models ongoing, some participants did point to problems in the funding of arXiv, a well-established preprint server, which had been through several periods of uncertainty with regard to its funding during its history.

Despite this, most participants seemed to favour a not-for-profit approach to preprint servers, some for reasons of sustainability, others emphasising the need for independence. Developments involving publishers becoming involved in delivering preprint services or in posting preprints on behalf of authors were, therefore, viewed with suspicion or even hostility by some participants. A number expressed concerns about the consolidation of services associated with academic workflows into the hands of a few commercial companies, including Elsevier ownership of SSRN, and Digital Science’s ownership of Figshare. This was seen as potentially jeopardising the independence of the relevant preprint services, even if they were currently operating in a standalone way. The language used by some participants to describe this development was in some cases strong, with one describing the sale of SSRN to Elsevier as “a huge betrayal of trust” (Engaged researcher) and another observing that this and similar developments have led to “power concentration” (Research performing organisation) in a small number of commercial providers. On the other hand, one representative from the preprint provider observed, “I don’t think that your average researcher thinks about that” (Preprint server provider); and it was indeed evident that many researchers amongst our participants did not show an awareness of such issues.

The future of preprints

There was considerable uncertainty amongst participants about the future role of preprint servers. When questioned whether preprint servers could form a significant part of a system of scholarly communication which would be an alternative to or replace peer-reviewed journals, most were sceptical. There was, however, some discussion of the potential value of ‘overlay journals’ – where virtual journals are created from content held in preprint servers, having been peer-reviewed and selected after their circulation. Some suggested that automated filtering rather than human-based peer review might have a role to play in creating overlay services. There was some awareness of experimental work in the overlay area, but few were able to identify working examples.

It was clear from comments of participants that if preprints were to play a more significant role in scholarly communication, major improvements to the preprints infrastructure would be needed. This would include incorporation of preprints into scholarly and publisher workflows, provision for production of preprints in standards-based formats (e.g. XML) and greater consideration of preservation services. All of this would require major investment. However, even in a system where preprints did not replace existing channels of communication, such as journals, many of these developments were still considered necessary in order to make the preprint infrastructure more robust.

Whilst uptake of preprints was seen by many to be increasing, the role of policy in this area was uncertain. It was seen as particularly problematical for use of preprints to be mandated, as opposed to encouraged, by funders. There was a clear perception that preprints should be adopted by researchers who see the benefits themselves, rather than in response to a mandate:

  • “There needs to be an intrinsic interest of the research community to communicate via preprints. I don’t think preprint posting can be enforced top-down or from anyone other than the research community and specifically the disciplinary communities themselves.” (Research funder)

It is noteworthy that this view was expressed by a research funder.

Discussion and conclusions

The findings of this research clearly relate to the innovation adoption decision process identified in Innovation Diffusion Theory, beginning with developing knowledge of innovation to confirmation of its adoption ( Table 7).

Table 7. Findings in relation to innovation adoption decision making factors from innovation diffusion theory.

Decision making
step
Details – when the decision maker/
making unit…
Summary of current findings
1. Knowledge…is exposed to the innovation’s
existence and gains an understanding
of how it functions
All interviewees chosen were already aware of preprints. However, we
note that there is uncertainty around what ‘the rules of the game’ are: while
people may be aware of preprints, the extent to which they are familiar
with their value proposition varies.
2. Persuasion…forms a favourable or unfavourable
attitude towards the innovation
For some, the hypothetical advantages are clear, but they find it difficult to
identify examples. Occasionally, there are doubts around whether posting
a preprint may limit chances of publishing. Research institutions often do
not have sufficient time and resources to promote and support preprints.
In our sample, more interviewees showed a favourable attitude than an
unfavourable one.
3. Decision…engages in activities that lead to a
choice to adopt or reject the innovation
In some cases, the Decision stage is strongly affected by the behaviour
of peers, e.g. a co-author wishing to post a preprint. In other cases,
the decision to post is supported by a belief in open scholarship and
transparency. The failure to adopt the posting and/or reusing of preprints
is often due to lack of uptake within a disciplinary community. In any event,
‘trialability’ is important
before a decision is made.
4. Implementation…puts an innovation into useExperimentation was mentioned often in our interviews. It applies to
preprint servers and overlay platforms in the first place, but also to
some researchers who are trying to establish whether preprints can be
beneficial to them – this is also related to the idea of trialability.
5. Confirmation….seeks reinforcement for an
innovation-decision already made but
may reverse the decision if exposed to
conflicting messages about it
Some interviewees reported that their or someone else’s preprints gained
attention and feedback, particularly on Twitter. Feedback was sometimes
received from important people within their disciplinary communities (e.g.
researchers and editors), which may lead to an improved article, new
connections or publication in prestigious journals.

It was evident that our participants were at different stages in the adoption process and this is a reflection of their peers’ and subject communities’ practices. Knowledge of preprints is rising but many are still not beyond the persuasion stage. Community norms remain crucial and have not shifted in many cases, therefore constraining individuals’ decisions. There was, nevertheless, some willingness to experiment, particularly amongst general OA supporters. There was some awareness of potential benefits becoming evident in practice but still at low levels. There was evidence of incomplete knowledge or misunderstandings amongst some researchers relating to preprints. The rate of adoption is influenced in our data by a number of factors highlighted in IDT ( Table 8).

Table 8. Findings in relation to the rate of innovation adoption from innovation diffusion theory.

VariableComponentsSummary of current findings
Perceived
attributes
Relative advantageThe potential value of preprints is clear to most although the existence of a relative
advantage for their use is not.
CompatibilityThe compatibility of preprints with existing practices and systems is perceived as variable
and often unclear.
ComplexityPerceived complexity of publisher policies and additional workload associated with deposit
can be a disincentive.
TrialabilityPreprint servers can be a strong basis for experimentation.
ObservabilitySome participants have observed the benefits of preprints based on colleagues’
experiences.
Communication
channels
Mass mediaThe widespread use of Twitter is a noticeable characteristic which enables discoverability
of preprints and growing awareness of preprint servers. It should be noted, however, that
use of Twitter is not universal e.g. in China.
InterpersonalPersonal recommendations are crucial in encouraging uptake, but were still often not
present.
Nature of the social
system
NormsPreprints are dealt with differently based on whether people are early adopters of
open science practices. In most other cases, preprints are considered as an important
development, but scepticism still has to be overcome (e.g. with respect to practical
advantages, funding streams and long-term preservation). Disciplines also play a
significant role, as preprints are widespread in some research areas but only emerging in
many others.
InterconnectednessCommunities are important in shaping perceptions of the value of preprints. We found that
two factors related to interconnectedness affect attitudes towards preprint posting: (i) the
adjacency to disciplines where preprint posting is common (e.g. disciplines close to others
within the scope of arXiv); and (ii) pressures and attitudes of close peers, particularly co-
authors.
Extent of change
agents’ promotion
efforts
Promotion efforts are limited, and it is unclear whose role this should be. Stakeholders in the
open science arena are promoting preprints within their circles and online. Some funders
(e.g. EC, Wellcome) are making explicit efforts to promote preprints, but more significant
and broader support (including from publishers) is likely to be required for higher uptake.
Type of innovation
decision
Optional, collective
or authority
Likely to continue to be optional for individuals influenced by their disciplinary community.
Mandated uptake is seen as unlikely although encouragement from policy makers is likely
to be strengthened.

It is worth mentioning that, in some cases, it is not clear whether researchers do not post preprints because there is no discipline-appropriate server for them, or there is no server because researchers in the field do not (want to) post them. The data shows that the willingness to be an early adopter of preprints may be related to sympathy with general open science and open access goals, as well as to the potential benefits researchers may see in their own use of preprints. Others remain sceptical, fearing that journals may reject their submissions, which is still the case in some areas ( Mallapaty, 2019), and questioning the value of circulation of pre-peer-reviewed outputs. The current environment for many disciplinary communities is therefore characterised currently by some experimentation, but also by uncertainty and fragmentation.

A key issue is trust. Trust is an essential feature of scholarly communication and was a recurring theme in the data, along with responsible posting and use of preprints. Nicholas et al. (2014) have shown that in the context of peer-reviewed journals, “researchers play down difficulties of establishing trustworthiness, not because there are none, but because they have well-developed methods of establishing trust”. However, preprints cut across those methods and create new ambiguities and uncertainties. The concern about lack of quality control and lack of quality indicators associated with preprints is fundamentally a matter of trust in a context where the “well-developed methods of establishing trust” are no longer present. Many of our participants were conscious that new norms were needed ( COPE Council, 2018) in this space and it was often the apparent absence of these that limited enthusiasm for preprints.

At a systemic level, the issue of sustainability emerged as critical from our work. Four models for delivering preprint services emerged from the data, each of which has different implications for sustainability and funding, an issue which was emphasised as being a concern by many participants:

  • 1.

    Standalone preprint servers using in-house technologies e.g., arXiv;

  • 2.

    Standalone preprint servers using third-party technologies e.g. bioRxiv using Highwire, ChemRxiv using Figshare infrastructure;

  • 3.

    Publisher-supported preprints e.g. PeerJ Preprints (now closed), F1000; and

  • 4.

    Publisher posting preprints to a preprint server e.g. PLOS partnership with bioRxiv.

Models 1 and 2 involve an ‘author-driven’ mode of posting preprints; Models 3 and 4 are a ‘publisher driven’ mode.

Model 1 is the ‘classic’ model of preprints, apparently assumed by most participants in their interviews. Model 2 is a version of 1 in which some infrastructure is outsourced to a third party, something which might help to enable sustainability by creating efficiencies through economies of scale (a common benefit of outsourcing). The infrastructure provided by the Center for Open Science is an example of such a benefit, with multiple preprint servers being run by the same organisation and on the same infrastructure, avoiding duplication, and therefore creating efficiency. Models 1 and 2 also include publisher-operated services such as Preprints.org and SSRN, which are owned by MDPI and Elsevier, respectively. These are not classified as publisher-supported (Models 3 and 4) but as standalone servers, because they still follow a paradigm of individual authors being responsible for posting their own preprints.

In this respect, Models 3 and 4 are different, since they involve the publisher rather than the author driving the preprint submission process, either through the publisher itself providing a preprints service as part of a journal submission workflow or by depositing on the author’s behalf. Models 3 and 4 move away from the traditional ‘author-driven’ preprints practices – where the author voluntarily deposits a paper as part of their own workflow, separately from submission to a journal – to a ‘publisher-driven’ preprints model, where the publisher to whom a manuscript has been submitted makes it available as a preprint. This is a fundamental shift which has major implications for the way the role of preprints is understood and the way preprints services are configured.

The future of preprints servers and their links with the overall scholarly communication process and infrastructure remain unclear. It is possible that the recent rise in preprint services might be reversed and that preprints go through a period of retrenchment, returning to serve the core areas traditionally associated with preprint use, such as high-energy physics. A possible alternative to such a ‘retrenchment scenario’ is what might be called the ‘patchiness scenario’, where different levels of adoption exist across different fields. Patchiness may be an ongoing situation, or may be a transition stage towards a possible ‘ubiquity scenario’. For preprints to become ubiquitous, of course, requires significant cultural and infrastructural change, some of which is indicated by the data presented here. This may partly depend on the integration between preprint services and other parts of the scholarly communication infrastructure and on related cultural norms. Closer integration may give rise to the possibility of more radical change in scholarly communication, creating opportunity for developments, such as overlay journals. Overlay journals have been discussed as thought experiments since at least the turn of the 21 st Century ( Smith, 2000; Smith, 1999) and there have been notable experiments in this area, and some ongoing services in specialised areas do exist, such as Discrete Analysis ( Ball, 2015). However, we are yet to see their widespread adoption, even though the potential remains.

Currently, the preprints landscape is rapidly changing, and disciplinary communities are at different stages in the innovation diffusion process. The high level of experimentation means that a one-size-fits-all approach to preprints is neither feasible or appropriate at present (if it ever will be). Something that will clearly play a role in the success of preprints (or lack thereof) is cooperation between the range of stakeholders involved. Though some of the issues we highlighted might appear independent of one another, we note that the majority of these affect multiple stakeholders at once.

Our findings have given rise to a number of key questions that we believe need to be addressed so that preprints can be supported sustainably in the future. Active engagement with these questions should lead to improved clarity and provide solid foundations for policy development and implementation. The questions and their relationships to the different actors involved are illustrated in Figure 1 (a tabular version of Figure 1 is available in Chiarelli et al. (2019c)). These include some key questions that need to be addressed by particular stakeholder groups, such as funders or preprint server providers. We have also added publishers as a separate group to this figure, taking account of issues that have arisen from the groups considered in this current study. However, there are also a large number of questions that need to be addressed through dialogue between different stakeholder groups. These are illustrated in the figure. One of the key challenges is to find channels for such dialogue to take place in order to develop solutions which are widely accepted.

Figure 1. Questions to address in the preprints landscape ( Chiarelli et al., 2019d).

Figure 1.

This is in many respects an agenda for further research, discussion and policy design. We expect, following a rapid rise in preprints since about 2013, that many of these questions will come to be seen as increasingly important over the next five years. The urgency with which they are addressed, and the ways in which they are answered by the different stakeholders, will shape the role that preprints play in scholarly communications in the future.

Data availability

Underlying data

The authors confirm that, for approved reasons, some access restrictions apply to the data underlying the findings. Data underlying this study cannot be made publicly available in order to safeguard participant anonymity and that of their organisations. Ethical approval for the project was granted on the basis that only extracts of interviews would be shared (with appropriate anonymisation) as part of publications and other research outputs. In order to share data with other researchers, the participants must be contacted and consent to this data being released. In order to request data release, other researchers should contact the corresponding author or Chair of the University of Sheffield Information School Research Ethics Committee ( ischool_ethics@sheffield.ac.uk).

Extended data

Zenodo: Interviewees and mapping of interview questions to areas of Innovation Diffusion Theory, http://doi.org/10.5281/zenodo.3538919 ( Chiarelli et al., 2019a).

Extended data.csv contains our interview questions, split by stakeholder group and mapped to innovation diffusion theory. List of interviewees.csv contains the names, roles and affiliations of interviewees.

Extended data are available under the terms of the Creative Commons Zero “No rights reserved” data waiver (CC0 1.0 Public domain dedication).

Acknowledgements

This article draws on the findings of an initial phase of work on the preprints landscape ( Chiarelli et al., 2019b) supported by the KE Task and Finish Group on preprints.

Funding Statement

This study has been funded by Knowledge Exchange (KE), a group of national organisations from six European countries supporting research infrastructure and services to enable the use of digital technologies to improve higher education and research: CSC in Finland, CNRS in France, DEFF in Denmark, DFG in Germany, Jisc in the UK and SURF in the Netherlands.

[version 2; peer review: 3 approved

References

  1. Abdill RJ, Blekhman R: Tracking the popularity and outcomes of all bioRxiv preprints. eLife. 2019;8: pii: e45133. 10.7554/eLife.45133 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. ASAPbio: University policies and statements on hiring, promotion, and journal license negotiation. (n.d.). Reference Source [Google Scholar]
  3. Balaji B, Dhanamjaya M: Preprints in Scholarly Communication: Re-Imagining Metrics and Infrastructures. Publications. 2019;7(1):6 10.3390/publications7010006 [DOI] [Google Scholar]
  4. Ball P: Leading mathematician launches arXiv "overlay" journal. Nature. 2015[cited 2019 Oct 11];526(7571):146 10.1038/nature.2015.18351 [DOI] [Google Scholar]
  5. Baughman S, Roebuck G, Arlitsch K: Reporting practices of institutional repositories: Analysis of responses from two surveys. J Libr Adm. 2018;58(1):65–80. 10.1080/01930826.2017.1399705 [DOI] [Google Scholar]
  6. Berg JM, Bhalla N, Bourne PE, et al. : SCIENTIFIC COMMUNITY. Preprints for the life sciences. Science. 2016;352(6288):899–901. 10.1126/science.aaf9133 [DOI] [PubMed] [Google Scholar]
  7. bioRxiv. Reference Source [Google Scholar]
  8. Bourne PE, Polka JK, Vale RD, et al. : Ten simple rules to consider regarding preprint submission. PLoS Comput Biol. 2017;13(5):e1005473. 10.1371/journal.pcbi.1005473 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Braun V, Clarke V: Using thematic analysis in psychology. Qual Res Psychol. 2006;3(2):77–101. 10.1191/1478088706qp063oa [DOI] [Google Scholar]
  10. Callaway E: Preprints come to life. Nature. 2013;503(7475):180. 10.1038/503180a [DOI] [PubMed] [Google Scholar]
  11. Cambridge University Press: Cambridge announces open research platform.2019. Reference Source [Google Scholar]
  12. Carà PD, Ciriminna R, Pagliaro M: Has the time come for preprints in chemistry? ACS Omega. 2017;2(11):7923–7928. 10.1021/acsomega.7b01190 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Carneiro CFD, Queiroz VGS, Moulin TC, et al. : Comparing quality of reporting between preprints and peer-reviewed articles in the biomedical literature. BioRxiv. 2019. 10.1101/581892 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Chiarelli A, Johnson R, Pinfield S, et al. : Interviewees and mapping of interview questions to areas of Innovation Diffusion Theory.2019a. 10.5281/zenodo.3538919 [DOI]
  15. Chiarelli A, Johnson R, Pinfield S, et al. : Practices, drivers and impediments in the use of preprints (Phase 1 report).2019b. 10.5281/ZENODO.2654832 [DOI] [Google Scholar]
  16. Chiarelli A, Johnson R, Pinfield S, et al. : Questions for future developments in the preprints landscape.2019c. 10.5281/zenodo.3539032 [DOI] [Google Scholar]
  17. Chiarelli A, Johnson R, Pinfield S, et al. : Accelerating scholarly communication: The transformative role of preprints.2019d. 10.5281/zenodo.3357727 [DOI] [Google Scholar]
  18. Chodacki J, Lemberger T, Lin J, et al. : Technical aspects of preprint services in the life sciences: a workshop report. Res Ideas Outcomes. 2017;3:e11825 10.3897/rio.3.e11825 [DOI] [Google Scholar]
  19. Cobb M: The prehistory of biology preprints: A forgotten experiment from the 1960s. PLoS Biol. 2017;15(11):e2003995. 10.1371/journal.pbio.2003995 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. COPE Council: COPE Discussion document: Preprints.2018. Reference Source [Google Scholar]
  21. Crossref: Crossref to accept preprints in change to long-standing policy | EurekAlert! Science News.2016; Retrieved May 15, 2019. Reference Source [Google Scholar]
  22. Davis PM, Fromerth MJ: Does the arXiv lead to higher citations and reduced publisher downloads for mathematics articles? Scientometrics. 2007;71(2):203–215. 10.1007/s11192-007-1661-8 [DOI] [Google Scholar]
  23. Desjardins-Proulx P, White EP, Adamson JJ, et al. : The case for open preprints in biology. PLoS Biol. 2013;11(5):e1001563. 10.1371/journal.pbio.1001563 [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Fraser J, Polka J: Together scientists and journalists can spot poor preprints. Nature. 2018;560(7720):553. 10.1038/d41586-018-06053-5 [DOI] [PubMed] [Google Scholar]
  25. Fraser N, Momeni F, Mayr P, et al. : The effect of bioRxiv preprints on citations and altmetrics. BioRxiv. 2019. 10.1101/673665 [DOI] [Google Scholar]
  26. Fry NK, Marshall H, Mellins-Cohen T: In praise of preprints. Microb Genom. 2019;5(4): e000259. 10.1099/mgen.0.000259 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Fu DY, Hughey JJ: Releasing a preprint is associated with more attention and citations for the peer-reviewed article. BioRxiv. 2019. 10.1101/699652 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Ginsparg P: Preprint Déjà Vu. EMBO J. 2016;35(24):2620–2625. 10.15252/embj.201695531 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Green T: Is open access affordable? Why current models do not work and why we need internet-era transformation of scholarly communications. Learn Publ. 2019;32(1):13–25. 10.1002/leap.1219 [DOI] [Google Scholar]
  30. Hampson C: The adoption of open access funds among Canadian academic research libraries, 2008-2012. Partnership: The Canadian Journal of Library and Information Practice and Research. 2014;9(2):2008–2012. 10.21083/partnership.v9i2.3115 [DOI] [Google Scholar]
  31. Harnad S: The Invisible Hand of Peer Review. Nature. 1998. Reference Source [Google Scholar]
  32. Johansson MA, Reich NG, Meyers LA, et al. : Preprints: An underutilized mechanism to accelerate outbreak science. PLoS Med. 2018;15(4):e1002549. 10.1371/journal.pmed.1002549 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Jones R, Andrew T, MacColl J: The institutional repository. Oxford: Chandos Publishing.2006. Reference Source [Google Scholar]
  34. Kaiser J: The preprint dilemma. Science. 2017;357(6358):1344–1349. 10.1126/science.357.6358.1344 [DOI] [PubMed] [Google Scholar]
  35. Kelly AR, Autry MK: Access, accommodation, and science: Knowledge in an "open" world. First Monday. 2013;18(6). 10.5210/fm.v18i6.4341 [DOI] [Google Scholar]
  36. Khera R: Sharing science at today’s pace: an experience with preprints. BMJ. 2018. Reference Source [Google Scholar]
  37. Klein M, Broadwell P, Farb SE, et al. : Comparing published scientific journal articles to their pre-print versions.2016;1–16. [Google Scholar]
  38. Kleinert S, Horton R, Editors of the Lancet family of journals: Preprints with The Lancet: joining online research discussion platforms. Lancet. 2018;391(10139):2482–2483. 10.1016/S0140-6736(18)31125-5 [DOI] [PubMed] [Google Scholar]
  39. Kling R, McKim G: Not just a matter of time: Field differences and the shaping of electronic media in supporting scientific communication. J Am Soc Inf Sci. 2000;51(14):1306–1320. [DOI] [Google Scholar]
  40. Krumholz HM, Ross JS, Otto CM: Will research preprints improve healthcare for patients? BMJ. 2018;362:k3628. 10.1136/bmj.k3628 [DOI] [PubMed] [Google Scholar]
  41. Larivière V, Sugimoto CR, Macaluso B, et al. : arXiv E-prints and the journal of record: An analysis of roles and relationships. J Assoc Inf Sci Technol. 2014;65(6):1157–1169. 10.1002/asi.23044 [DOI] [Google Scholar]
  42. Lauer MS, Krumholz HM, Topol EJ: Time for a prepublication culture in clinical research? Lancet. 2015;386(10012):2447–2449. 10.1016/S0140-6736(15)01177-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Lee CJ, Sugimoto CR, Zhang G, et al. : Bias in peer review. J Am Soc Inf Sci Technol. 2013;64(1):2–17. 10.1002/asi.22784 [DOI] [Google Scholar]
  44. Leopold SS, Haddad FS, Sandell LJ, et al. : Editorial: Clinical Orthopaedics and Related Research, The Bone & Joint Journal, The Journal of Orthopaedic Research, and The Journal of Bone and Joint Surgery Will Not Accept Clinical Research Manuscripts Previously Posted to Preprint Servers. Clin Orthop Relat Res. 2019;477(1):1–4. 10.1097/CORR.0000000000000565 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Luther J: The stars are aligning for preprints. Retrieved December 11, 2017, from Scholarly Kitchen [web blog],2017. Reference Source [Google Scholar]
  46. Mallapaty S: Indian scientists launch preprint repository to boost research quality. Nature. 2019. 10.1038/d41586-019-01082-0 [DOI] [PubMed] [Google Scholar]
  47. Martín-Martín A, Costas R, van Leeuwen T, et al. : Evidence of open access of scientific publications in Google Scholar: A large-scale analysis. J Informetr. 2018;12(3):819–841. 10.17605/OSF.IO/K54UV [DOI] [Google Scholar]
  48. McGlynn T: What’s up with preprints? Retrieved November 14, 2019, from Small Pond Science. Published 2017. Reference Source [Google Scholar]
  49. McKiernan EC, Schimanski LA, Nieves CM, et al. : Use of the Journal Impact Factor in academic review, promotion, and tenure evaluations.2019. 10.7287/peerj.preprints.27638v2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Nallamothu BK, Hill JA: Preprints and Cardiovascular Science: Prescient or Premature? Circ Cardiovasc Qual Outcomes. 2017;10(9):pii: e000033. 10.1161/HCQ.0000000000000033 [DOI] [PubMed] [Google Scholar]
  51. Nature Preceedings. Reference Source [Google Scholar]
  52. Neylon C, Pattinson D, Bilder G, et al. : On the origin of nonequivalent states: How we can talk about preprints [version 1; peer review: 2 approved]. F1000Res. 2017;6:608. 10.12688/f1000research.11408.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Nicholas D, Watkinson A, Volentine R, et al. : Trust and authority in scholarly communications in the light of the digital transition: Setting the scene for a major study. Learn Publ. 2014;27(2):121–134. 10.1087/20140206 [DOI] [Google Scholar]
  54. PeerJ: PeerJ preprints to stop accepting new preprints Sep 30th 2019. PeerJblog2019. Reference Source [Google Scholar]
  55. Pinfield S: Self-archiving publications. In G. Gorman & F. Rowland (Eds.), Scholarly Publishing in and Electronic Era: International Yearbook of Library and Information Management 2004-2005 London: Facet Publishing.2004;118–145. Reference Source [Google Scholar]
  56. Pinfield S, Middleton C: Researchers’ adoption of an institutional central fund for open-access article-processing charges: A case study using Innovation Diffusion Theory. SAGE Open. 2016;6(1). 10.1177/2158244015625447 [DOI] [Google Scholar]
  57. Polka J: Preprints in academic hiring. Retrieved March 13, 2019, from DORA blog,2018. Reference Source [Google Scholar]
  58. Research England: REF 2019/01: Guidance on submissions. London;2019. Reference Source [Google Scholar]
  59. Rittman M: Preprints as a Hub for Early-Stage Research Outputs.2018. 10.20944/preprints201806.0243.v1 [DOI] [Google Scholar]
  60. Rogers EM: Diffusion of innovations. (5th ed.). New York: The Free Press.2003. Reference Source [Google Scholar]
  61. Sarabipour S, Debat HJ, Emmott E, et al. : On the value of preprints: An early career researcher perspective. PLoS Biol. 2019;17(2):e3000151. 10.1371/journal.pbio.3000151 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Sarabipour S: Preprints are good for science and good for the public.2018. Reference Source [DOI] [PubMed] [Google Scholar]
  63. Sever R, Eisen M, Inglis J: Plan U: Universal access to scientific and medical research via funder preprint mandates. PLoS Biol. 2019;17(6):e3000273. 10.1371/journal.pbio.3000273 [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Sheldon T: Preprints could promote confusion and distortion. Nature. 2018;559(7715):445. 10.1038/d41586-018-05789-4 [DOI] [PubMed] [Google Scholar]
  65. SHERPA/RoMEO: SHERPA RoMEO Colours, Pre-print, Post-print, Definitions and Terms. Retrieved May 15, 2019. (n.d.). Reference Source [Google Scholar]
  66. Smith A: The journal as an overlay on preprint databases. Learn Publ. 2000;13(1):43–8. 10.1087/09531510050145542 [DOI] [Google Scholar]
  67. Smith JWT: The deconstructed journal - a new model for academic publishing. Learn Publ. 1999;12(2):79–91. 10.1087/09531519950145896 [DOI] [Google Scholar]
  68. Suber P: Open Access. MIT Press.2012. Reference Source [Google Scholar]
  69. Teixeira da Silva JA: The preprint wars. AME Med J. 2017;2:74 10.21037/amj.2017.05.23 [DOI] [Google Scholar]
  70. Teixeira da Silva JA, Dobránszki J: Preprint policies among 14 academic publishers. J Acad Libr. 2019;45(2):162–170. 10.1016/j.acalib.2019.02.009 [DOI] [Google Scholar]
  71. Tennant J, Bauin S, James S, et al. : The evolving preprint landscape: Introductory report for the Knowledge Exchange working group on preprints. BITSS. 2018;5 10.17605/OSF.IO/796TU [DOI] [Google Scholar]
  72. Tennant J, Gatto L, Logan C: Preprints help journalism, not hinder it. Nature. 2018;560(7720):553. 10.1038/d41586-018-06055-3 [DOI] [PubMed] [Google Scholar]
  73. Tennant J, Crane H, Crick T, et al. : Ten hot topics around scholarly publishing. Publications. 2019;7(2):34 10.3390/publications7020034 [DOI] [Google Scholar]
  74. Twitter. Reference Source [Google Scholar]
  75. Vale RD: Accelerating scientific publication in biology. Proc Natl Acad Sci U S A. 2015;112(44):13439–13446. 10.1073/pnas.1511912112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Velterop J: Of subscriptions and article processing charges.2018. Reference Source [Google Scholar]
  77. Wang X, Cui Y, Xu S, et al. : The state and evolution of Gold open access: a country and discipline level analysis. Aslib J Inform Manag. 2018;70(5):573–584. 10.1108/AJIM-02-2018-0023 [DOI] [Google Scholar]
  78. Warr WA: Evaluation of an experimental chemistry preprint server. J Chem Inf Comput Sci. 2003;43(2):362–373. 10.1021/ci025627a [DOI] [PubMed] [Google Scholar]
  79. Xia J: Diffusionism and open access. J Doc. 2012;68(1):72–99. 10.1108/00220411211200338 [DOI] [Google Scholar]
F1000Res. 2020 Feb 4. doi: 10.5256/f1000research.23536.r58477

Reviewer response for version 2

Todd Carpenter1

This review of “Preprints and Scholarly Communication: Adoption, Practices, Drivers and Barriers,” covers the work by Chiarelli et al. to survey a group of 38 stakeholders in the preprint and scholarly community in a limited set of scientific domains, specifically, biology, chemistry and psychology.

Methodology:

As a framework for assessing the adoption of Preprints, the Rogers model is a solid one, and the authors do a good job of crafting a survey around the five core elements of the basic conceptual model.

One gap, and I feel it is an important one to be explicit about, particularly regarding Rogers’ conceptualization is the lack of focus on the discussion of motivations. Rogers defines this as the Knowledge-Attitude-Practice (KAP) Gap. To summarize, there is always a cost to make a change in any technical adoption process. The individual must be sufficiently motivated to make the change and thereby overcome the costs of making the change and adopting the new technology. With regards to motivations for adoption of preprints, this is a core element of the drivers of adoption, perhaps more so than the variables and findings outlined in Table 8 (which do follow Rogers’ approach). The potential ‘costs’ related to reputational damage, dissemination of erroneous results, intellectual property theft, or being blocked from formal publication. These costs change depending on the field and could correlate strongly with the willingness to use preprint systems.

In table 2, the authors outline variables affecting the rate of adoption. Missing from this, based on KAP-Gap element of Rogers is the concept of the “conservative-ness” of the academic marketplace, where the risks and implications of failure can be significant. These risks increase in some academic domains, but remain regardless of the field. There is also a strong fixity and lack of receptivity of change in the academic social system which also inhibit adoption based on increased levels of ‘cost’ in the Rogers model.

While this is not a critical flaw in the paper, focusing on this added element would draw out many of the themes that appear in a variety of the sections and give focus to the very real barriers that exist in the preprint marketplace. One can see that the authors skirt around this issue in some of their questioning and in their analysis. A stronger methodological framework would have highlighted these themes better.

It seems that the interviews were intentionally selected to focus on particular domains and to include people with specific engagement with particular communities. While useful in a knowledge gathering process for a qualitative survey and the paper does a reasonable job justifying its selection rationale (page 8 & 9), the actual selection of participants seemed to be intentional, and this does add the concern about selection bias.

References:

While I did not review the earlier draft, it appears that the literature review is reasonably balanced and covers a significant swath of the relevant literature, though to be fair this isn’t a review article and I don’t expect it to be comprehensive of every paper in this space.

Results: 

There is much within the results section to dig into intellectually and the feedback is useful qualitative reporting of the opinions of the subjects. Again, with the caveat that there is a potential for selection bias in these results, several additional details would improve the paper.

Repeatedly, there references to approximations of the breadth of opinions expressed in the paper, such as page 10 “a small number of interviewees, or in table 5, where quantification of these amounts might be informative. I do understand that this is not a quantitative study and that including those figures might give that impression.

There were inconsistencies or contradictions in a number of the responses, which are referenced in the reporting of the results, which is an indication of the variety of responses and perspectives covered in this survey. Some examples of these contradictions include the trust in preprints as a component of the publication workflow in which the manuscript would eventually be published, but yet preprint servers are also reported as a home for “orphaned” works that can’t find a ‘traditional’ publication home (page 13). The core value of preprints as a home for open access content delivery, but there is a rapidly growing acceptance of open access publication in the traditional space (page 13). That the readers of preprints should accept greater responsibility for assessing the quality of the papers, while also promoting their value as a source of information for the broader public, who likely are least well-positioned to assess the quality of a research output (page 14). The tension between the need for quality and trusted review of science and the desire to get results out the door as quickly as possible (page 14/15). This last tension is highlighted in the current environment by the rapid availability of manuscripts posted in preprint repositories on the coronavirus first identified in Wuhan, Hubei Province, China in the winter of 2020.

I would have hoped for a fuller discussion and consideration of these results in the conclusion section of the paper. Again, I do not find this to be a fatal flaw of the paper. Perhaps the publication of these results will add to the urgency for a need to discuss these tensions of opinions related to preprints.

The chicken-or-egg problem in the discussion section on whether the lack of repositories for a domain is limiting preprint deposit or the lack of demand is limiting creation of repositories. It seems in the current environment that much like there’s a journal for every paper, there is a repository for every preprint. This also seems out of place when the domains surveyed are ones in which specifically selected as having available repositories.

Overall, the paper is well-written and well-reported. I recommend approval for indexing.

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

F1000Res. 2020 Jan 13. doi: 10.5256/f1000research.23536.r58424

Reviewer response for version 2

Hamid Jamali1

This is an insightful and interesting article that reports the results of a qualitative study that includes interviews with key actors in the preprint domain in order to find out about the current and future role of the preprints. The study uses innovation diffusion theory as a framework. The paper is well-written and presents the results as it is the norm in social sciences. The literature review covers several aspects related to preprint that are relevant to the context of the research. The method has been described clearly and sufficiently. The findings are supported with evidence from the interviews (quotes) and they are discussed using the framework of the study (IDT). I found the comparison of the findings from the interviews with the literature and the questions for future studies (Figure 1) particularly useful.

I also enjoyed reading the comments of previous reviewers and the authors’ response to those comments. I believe the methodology is fine and well-justified in the current version.

I have no more corrections to suggest.

As a minor point, RPO used in the figure hasn’t been mentioned in the paper and I had to find out what it stands for, although the phrase has been used in the paper.

I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

F1000Res. 2019 Dec 2. doi: 10.5256/f1000research.23536.r57003

Reviewer response for version 2

Jessica K Polka1, Naomi C Penfold1

The authors' additions greatly enrich and clarify this version of the manuscript. While our queries were addressed, the additional description makes it clear that we lack the expertise to evaluate the methodology. We therefore recommend that a third reviewer with appropriate disciplinary expertise take a look at the revised manuscript.

We confirm that we have read this submission and believe that we have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however we have significant reservations, as outlined above.

F1000Res. 2019 Dec 2. doi: 10.5256/f1000research.23536.r57004

Reviewer response for version 2

Sarvenaz Sarabipour1, Humberto Debat2

The authors have adequately responded to our comments.

We confirm that we have read this submission and believe that we have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.

F1000Res. 2019 Aug 5. doi: 10.5256/f1000research.21512.r50950

Reviewer response for version 1

Jessica K Polka1, Naomi C Penfold1

In “Preprints and Scholarly Communication: Adoption, Practices, Drivers and Barriers,” Chiarelli and colleagues present a literature review, commentary, and a summary of attitudes and perceptions towards preprints found by interviewing 38 stakeholders. The interviews convey many commonly-held perceptions about the benefits and disadvantages of preprints, and the authors suggest questions for further exploration and testing. With revisions (clarifying the origin of claims, contextualizing the responses from study subjects, and correcting factual information in the text) this paper will become a very useful addition to the literature. 

Major comments

  • Design and Subjects:

     
    • We encourage authors to rely on a more robust review of published literature (including the grey literature) to complement their interviews, as the published literature contains useful context for comments which are obfuscated by the process of anonymous reporting (For example, the sentence, “One preprint service provider observed that only 10% of preprints received comments.” would be more useful to readers if accompanied by a reference with attribution: https://asapbio.org/biorxiv)
    • Since this work may be of interest to a wide range of readers, we recommend including more detail regarding the chosen experimental design. 
      • Please provide justification for the experimental design, for example, why were interviews used instead of a survey? 
      • What are the methodological considerations that would be helpful for readers outside of qualitative research to understand: limitations, constraints, advantages over more quantitative methods? 
      • Please define semi-structured interviews — how were researchers encouraged to expand on thoughts? 
      • Please define “purposive heterogeneous sampling approach.”
      • Please clarify why the sample size is 38.
    • Please provide more information about the subjects to help readers assess relevance:
      • What is their geographical context?
      • How knowledgeable were subjects, and do they bring bias for or against open science into their statements? 
      • How did interviewees arrive at their opinions? How much close experience do they have with preprints? Were participants asked to answer only within their own expertise? 
      • There are quoted statements that reflect inaccurate understandings and are not always qualified by the author: for example, “relating to whether preprints are seen as an appropriate object for evaluation in exercises such as the UK’s Research Excellence Framework, about which there remains ambiguity. One representative of a university stated: “they’re not acceptable for REF, so they’re not even part of the equation. So it’s the author’s accepted manuscript is the currency we deal with.” (Research performing organisation)”  This is inaccurate and misleading (see https://www.ref.ac.uk/publications/guidance-on-submissions-201901/) but there is no further commentary from the authors. Please add clarification that the interviewee responses are not necessarily factual. (It would be interesting to compare interviewee opinions with known facts to understand which topics are well-understood.)

 

  • Clarity of presentation:

     
    • Please provide more quantitative context to the opinions expressed. For example, for phrases such as “Some researchers” it would be useful to be more specific. Please also be more quantitative in assessment of the themes — it is useful to know if a sentiment was expressed by a majority or minority, as shown clearly in tables of benefits and challenges. Please provide aggregate data of thematic analyses, not just quotes, and use specific numbers and proportionality (Two of Five researchers mentioned X as a benefit).
    • The paper combines elements of a literature review, white paper/policy piece, and a qualitative study. It would be helpful to clarify the origin of ideas by separating the work clearly into Results and Discussion sections. (As one example, under the heading “Financial sustainability and business models,” it is unclear whether the proposed taxonomy was expressed by interviewees or generated by the authors. Similarly, it is unclear whether use of the term “seminal” comes from interviewees or authors.)
    • Readability:
      • Please use active voice and concise language to improve readability.
      • Bring interview questions and list of participants into this report — it is hard to interpret findings while having to continuously refer back to other sources, e.g. the 2019b interview structure paper.
  • Claims:

     
    • The abstract sets the reader up to expect a review of preprint servers that have emerged since 2013. However, the methods section discloses that coverage of the study is limited to biology, chemistry, and psychology. It would be helpful to readers to state this up front. It is also unclear how this scope was applied in the selection of study subjects. It is difficult to determine if the scope influenced the selection of participants as listed in https://zenodo.org/record/2654832#.XTeUB5NKjs0 – the questions listed at https://zenodo.org/record/3240426 do not seem specific to these three disciplines and the quoted answers in this report suggest interviewees’ responses were not limited to these three disciplines. Please clarify how the scope of the interviews was constrained to the stated subjects, or if not, please revise the stated scope to more accurately reflect the scope employed.
    • The abstract does not reflect the content. The interview responses are summarised in general terms, not related to defined scope or the newer servers. A substantial proportion of this article is review material: either in the explicit literature review section, or as author-contributed content interspersed with summaries of interviewee responses. For example, conclusions listed in the abstract appear to be drawn from the literature review.
    • “Our study is the first using empirical data to understand the new wave of preprint servers” — this is not supported by the study content: no results are specifically related to the newer servers identified in the introduction; quoted excerpts indicate that interviewees responded in general about all servers and disciplines. Furthermore, prior surveys and bibliometric analyses have looked at a subset of these servers, either individually or in groups (eg https://osf.io/5fk6c/). 1

Minor comments

We recommend making the following amendments to improve the clarity and accuracy of the report.

In the Introduction and literature review: 

  • The summary of the longer history of preprint efforts across disciplines, would benefit from incorporating and citing ‘The Brief Prehistory of Preprints” (Cobb, 2017) 2

  • Zuckerberg Foundation should be corrected to Chan Zuckerberg Initiative.

  • Discussion of new servers would be improved with a note as to their current size, and with awareness that not all new ‘preprint’ servers are exclusively for preprint content (many COS servers include postprints that are not possible to easily distinguish from preprint versions).

  • “This paper aims to explore…by investigating current practices, drivers and barriers to [preprint] use.” The interviews indicate perceptions and attitudes but do not reveal actual behaviours or practices. Please clarify how this study examines current practices.

Figures and tables:

  • Fig 1: To increase accessibility, please provide as raw text with headings or tags to indicate which stakeholder group you believe is most relevant

Literature review:

  • The authors focus on four main issues — how were these arrived at? Was the literature review systematic in any way? 

  • Given the emergent phase of preprint infrastructure in biology, chemistry and psychology, and as the authors note that empirical data is lacking, it may be equally valuable to include more grey literature (blogposts, webpages) in this review, given that several of the cited articles are opinion pieces, albeit as editorials or peer-reviewed review articles. Of note, we find personal blogs are a useful source of stories that detail benefits and drawbacks of preprints from personal experiences.

  • "Dis-benefits:" is not a common term, perhaps use “drawbacks,” “concerns,”, “disadvantages” or “potential risks”.

  • When discussing the lack of quality assurance, the authors seem to make an implicit assumption here that journal-led peer review assures quality. It would be fair to contextualise this section with some literature on peer review to understand variety of peer review outcomes on quality of paper and validity of findings.

Results:

  • In the discussion of preprints in the medical field, please note that the concerns raised are being addressed by medRxiv. https://www.medrxiv.org/

  • In the table of business models:

     
    • Note that bioRxiv uses 3rd party technology, Highwire.
    • “Publisher driven” models are not new; PeerJ is older than bioRxiv, Nature had Nature Preceedings in the early 2000s, and bioRxiv has had J2B preprints since its inception.
    • How does the outsourcing of infrastructure to a third party enable sustainability? Arguably, depending on the third party, this leaves the service vulnerable to changes in service entirely outside its control.
    • Individual authors can still submit to PeerJ.
    • Most publisher-mediated preprint submissions are voluntary for authors (a notable exception being F1000).
    • Regarding the statement of subject preferences: we have conducted a survey on these preferences that may provide additional context: https://asapbio.org/asapbio-funders-workshop 

The following suggestions may also help improve the paper but are not essential.

In the Introduction and literature review:

Figures and tables:

Literature:

  • “However, Bourne et al. (2017) controversially extend the definition to include “a paper that has been peer reviewed and…was rejected, but the authors are willing to make the content public”.” — Please provide reasoning or citation as to why this is controversial.

  • Accessibility — note that preprints are not necessarily open access, where they are not provided with permissive licensing, and this is certainly the case for bioRxiv. 3

  • One potential benefit to ECRs that is not mentioned here is being able to demonstrate productivity to funders and hirers. This would be useful to append to the current paragraph, given it is raised in the interview questions.

  • “There also appears to be little acknowledgement that the claim stands in tension to the one cited earlier that preprints often differ little from final published versions.” — There is only tension if both claims concern all preprints. It is possible for there to be preprints that improve through a lot of feedback as well as preprints that undergo no/very little change.

  • Please use unshared or unpublished instead of “homeless”.

  • We recommend avoiding the term “policy stack,” as there is no dependent relationship among these different groups/stakeholders.

Methods:

  • This section ignores the efforts of researchers as change agents.

Results:

  • Please providence evidence for this statement: “we note that adjacency between disciplines appears to be playing an important role.” 

  • “Many of the participants agreed that preprints servers could be a useful outlet for otherwise ‘homeless’ research outputs, even though this goes counter to the emphasis of many of preprints as early versions of outputs later formally published elsewhere.” — this attitude may reflect a lack of publication venues for such content, not a lack of desire to publish these other outputs in a peer-reviewed destination.

We confirm that we have read this submission and believe that we have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however we have significant reservations, as outlined above.

References

  • 1. : Quantifying the growth of preprint services hosted by the Center for Open Science.2019; 10.31219/osf.io/5fk6c 10.31219/osf.io/5fk6c [DOI] [Google Scholar]
  • 2. : The prehistory of biology preprints: A forgotten experiment from the 1960s. PLoS Biol.2017;15(11) : 10.1371/journal.pbio.2003995 e2003995 10.1371/journal.pbio.2003995 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. : Biologists debate how to license preprints. Nature.2017; 10.1038/nature.2017.22161 10.1038/nature.2017.22161 [DOI] [Google Scholar]
F1000Res. 2019 Nov 14.
Andrea Chiarelli1

In “Preprints and Scholarly Communication: Adoption, Practices, Drivers and Barriers,” Chiarelli and colleagues present a literature review, commentary, and a summary of attitudes and perceptions towards preprints found by interviewing 38 stakeholders. The interviews convey many commonly-held perceptions about the benefits and disadvantages of preprints, and the authors suggest questions for further exploration and testing. With revisions (clarifying the origin of claims, contextualizing the responses from study subjects, and correcting factual information in the text) this paper will become a very useful addition to the literature. 

Authors’ response:

We are grateful for the reviewers’ summary of and comments on our article, and are pleased that they consider that “with revisions…this paper will become a very useful addition to the literature.” The above summary makes a number of general remarks about revisions that should be made which we address below, responding to the detailed comments of the reviewers.

We believe that it may be helpful before addressing the specific comments below to precede our responses with an explanation of our approach to qualitative research in general. Many of the comments made by the reviewers in their report are as much about qualitative research in general as they are about our particular study, and so we would like to set out our approach in the round, as context to our detailed comments further below. The open peer review process may be helpful here, as the comments we make are openly available and help to explain our approach, even though not all of the detail can be incorporated in our final paper. This general explanation is reproduced in both our responses to reviewers’ comments, as similar issues arise in both.

Our general approach in responding to the comments of the reviewers on qualitative methods is to try to achieve a balance between, on the one hand, providing clarity and further detail where necessary, and, on the other hand, avoiding over-extending an already long paper. So we have made some changes to the text of the paper with the aim of achieving greater clarity but provide more detailed explanations here in in this report in order to provide fuller explanations that might be too detailed for the paper itself.

Our study uses well-established qualitative research methods. Bazeley (2013) states: "researchers engaging in a qualitative study focus on observing, describing, interpreting, and analysing the way that people experience, act on, or think about themselves and the world around them”. Qualitative research aims to collect and analyse “rich, deep data” (Bryman, 2015) which enable understandings of issues like motivations, beliefs, and values of participants in the contexts in which they act. It focuses on understanding the perspectives of participants, including where disagreement and conflict may exist. Qualitative research typically attempts not just to identify behaviours but also meanings associated with behaviours. It is often useful for encountering unexpected issues and influences. Qualitative methods are commonly deployed therefore in studying phenomena and contexts which are emerging, where issues need to be mapped out, and theory developed (which can then in turn potentially be tested in various ways, including quantitatively). We chose to explore the preprints space using these methods in order to gain a rich and deep understanding of the issues involved, specifically including varying motivations, differing behaviours, and conflicting perspectives.

Qualitative data can take different forms and be gathered in different ways, but often take the form of text, e.g. transcripts of interviews, as in our study. In this case the transcripts are the data being analysed, and there are well-established methods for carrying out such analysis. Analysis is typically conducted inductively, identifying patterns which emerge from the data. This is normally done through a process like thematic analysis (used in this study) involving a number of steps, as explained by Braun & Clarke (2006), who are cited in our study. There are a large number of other texts that describe this process, albeit with variations e.g. Bazeley (2013), Bryman (2015), Maxwell (2013) etc. In the first stage of the process, the interviews are fully transcribed and then read and reread thoroughly by the researchers. Secondly, the entire dataset is coded – key features of the data being labelled at a detailed level. “Codes…serve as shorthand devices to label, separate, compile and organize data” (Charmaz, 1983). This is an iterative process involving constant comparisons across the dataset. It can be carried out in various ways, but in our case, we used the NVivo software which facilitates coding of large amounts of text. Coding is a kind of fracturing process, breaking down the data into small fragments, and so the third stage of thematic analysis involves assembling codes into groups or themes. Themes are identified iteratively through constant comparisons, often involved in mapping out themes in relation to one another in order to define and refine them. These were the processes we followed in this study.

In reporting qualitative research, the themes are normally used as a framework, rather than the questions that were initially asked of participants. This is an important aspect of the reporting of an inductive study in order to represent the themes that emerged from the data rather than the initial framework of the questions asked. From analysis of these themes, theoretical insights should emerge. These could take the form of systematic explanations of a context; models of relationships, issues, challenges etc; sets of questions or hypotheses that need to be further investigated; etc. We have tried to provide some of these, albeit tentatively, in our study.

Sampling in qualitative research is normally carried out in a purposive (or purposeful) way. This is described by Bryman (2015):

“Purposive sampling is a non-probability form sampling. The researcher does not seek to sample research participants on a random basis. The goal of purposive sampling is to sample cases/participants in a strategic way, so that the sample are relevant to the research questions that are being posed. Very often, the researcher will want to sample in order to ensure that there is a good deal of variety in the resulting sample, so that sample members differ from each other in terms key characteristics relevant to the research question.”

We describe our sampling approach as a “heterogeneous purposive sampling approach”, which we go on to define as, “aiming to include a wide range of perspectives from actors in the area”. The sample was heterogeneous in a number of respects: it contained representatives of different roles in the scholarly communication system, from different countries (and therefore policy environments), and with different views and levels of experience of using preprints. Participants were selected by us based on an analysis of the field, consultation with our funders and others, and then approached directly and invited to participate. We also advertised on email discussion lists and social media inviting participants, but only selected people to be interviewed who met our sampling criteria. One challenge was in finding researchers willing to be interviewed who were not engaged with preprints, and so in order to do this, we not only used the selection method of approaching people directly but also used “snowball sampling” – asking for recommendations of other people who we might approach from participants. As Bryman (2015) comments:

“Purposive sampling often involves more than one of the approaches outlined… For example, it is quite common for snowball sampling to be preceded by another form of purposive sampling. In effect, the process entails sampling initial participants without using the snowball approach and then using these initial contacts to broaden out to a snowballing method.”

This is what we did, resulting in a final sample of 38.

Sample size is a controversial issue. There is no obvious right answer to the question of, “how many interviews are enough?” The essays addressing that question by well-known qualitative researchers edited by Baker & Edwards (2012) include a wide range of different views, depending on a number of factors. Bryman quotes Mason (2010) as having examined UK and Irish PhD theses using qualitative interviews, which identifies them as using a mean sample size of 31 and a median of 28. Bryman, in his chapter on reporting research, looks in detail at a sample journal article reporting semi-structured interviews which has a sample size of 20 (Jones, Leontowitsch, & Higgs, 2010). Our sample size of 38 is relatively large. One key point to emphasise, however, is that sample size in qualitative research is normally not determined in advance. Rather, the sampling approach will be decided and participants recruited within the sampling frame. The researchers will then make a number of judgements during the data gathering process about when sufficient data has been collected. Chief amongst these is the idea of “theoretical saturation”, where new data “no longer suggests new insights” (Bryman, 2015). This involves a number of judgements, involving a close engagement with the data in relation to the research objectives.

Because qualitative research usually involves identifying areas of motivations, perspectives etc, there is commonly a need to allow participant anonymity. Often interviewees are less frank when they are ‘on the record’. Less frank interviewees tends to lead to a rather bland dataset which makes it difficult to understand the real contours of a landscape. In our study, we gained ethical approval to name participants and their organisations (with their permission) but not to associate any particular quotations with individuals. We have adopted a conventional way of dealing with this by identifying various groups amongst our sample which we do relate with each quotation (“Unengaged researcher”, “Preprint service provider” etc). Naming the group from which the interviewee being quoted came helps to contextualise each quotation, but does not compromise promised anonymity.

Providing extensive extracts of participant quotations is a critical part of presenting findings from qualitative research. Since these quotations are illustrative of the themes identified, and are examples of the data collected, showing them in detail is an important way of demonstrating the validity of the inferences being drawn from the research.

Baker, S. E., & Edwards, R. (2012). How many qualitative interviews is enough? Retrieved from http://eprints.ncrm.ac.uk/2273/4/how_many_interviews.pdf

Bazeley, P. (2013). Qualitative data analysis: Practical strategies. London: Sage.

Bryman, A. (2015). Social research methods (5th ed.). Oxford: Oxford University Press.

Jones, I. R., Leontowitsch, M., & Higgs, P. (2010). The experience of retirement in second modernity. Sociology, 44(1), 103–120. https://doi.org/10.1177/0038038509351610

Maxwell, J. (2013). Designing a qualitative study. In L. Bickman & D. J. Rog (Eds.), The SAGE handbook of applied social research methods (pp. 214–253). https://doi.org/10.4135/9781483348858.n7

Major comments

Reviewers’ comments:

  • Design and Subjects:
    • We encourage authors to rely on a more robust review of published literature (including the grey literature) to complement their interviews, as the published literature contains useful context for comments which are obfuscated by the process of anonymous reporting (For example, the sentence, “One preprint service provider observed that only 10% of preprints received comments.” would be more useful to readers if accompanied by a reference with attribution: https://asapbio.org/biorxiv)

 

Authors’ response:

We have included a literature review in our study and are happy to update it with other relevant literature (as mentioned below and in response to comments of the other reviewers).

We sketch out the literature review and engage with literature again in our discussion, but limit our reporting of our data to doing exactly that – reporting our data. We report therefore that, “One preprint service provider observed that only 10% of preprints received comments” but we have to be careful not to appear to attribute this quotation in a way that makes the interviewee identifiable (see above comments on anonymity). As this quotation is from a “preprints service provider” it is reasonable to assume it is reliable.

Reviewers’ comments:

  • Since this work may be of interest to a wide range of readers, we recommend including more detail regarding the chosen experimental design. 
    • Please provide justification for the experimental design, for example, why were interviews used instead of a survey? 
    • What are the methodological considerations that would be helpful for readers outside of qualitative research to understand: limitations, constraints, advantages over more quantitative methods? 
    • Please define semi-structured interviews — how were researchers encouraged to expand on thoughts? 
    • Please define “purposive heterogeneous sampling approach.”
    • Please clarify why the sample size is 38.

Authors’ response:

To respond to the above bullet points in turn:

Justification for the approach. As the above general explanation outlines, we chose qualitative methods in order to explore in-depth the “varying motivations, differing behaviours and conflicting perspectives” of key actors in the preprints space – something qualitative methods are often designed to do. We have added some more detail to the method to clarify this, notably in the first paragraph of the Methods section:

“As preprints and preprint servers are still innovative developments for most disciplines, it is important to gain an in-depth understanding of the perspectives of different stakeholders. In order to explore issues, such as varying motivations, differing behaviours, and conflicting perspectives, particularly in emerging areas, qualitative research methods are often deployed, since they are well suited to such investigations. We chose to carry out detailed interviews of key actors in this space who could explain in depth their perceptions, attitudes and practices in relation to preprints…” etc.

Limitations and constraints. We have also added a new section in the Methods section on “Limitations and constraints”. The main limitation with qualitative research is lack of generalisability. Whilst results may be transferable to other contexts, generalisability is not possible from qualitative research alone. Qualitative research is designed to explore specific instances of phenomena and generate theory and hypotheses which might then be tested using other methods, including quantitative methods where findings might be generalised. Our new section reads:

“Limitations and constraints. Like many kinds of qualitative research, this study was designed to be exploratory; in this case, to map out key aspects of the preprints space and suggest policy responses.  Our conclusions are tentative. Many of interviewees were selected because of their knowledge of the issues under investigation, and although our findings based on their views may be transferable to other contexts, they cannot be generalised without further testing, as with most qualitative research. There was a bias in our sampling of participants aware of and engaged with preprints. Further research, using other methods, will be needed in future in order to generalise across communities as a whole, including non-engaged researchers. Furthermore, some stakeholder groups, such as publishers (who only have very limited representation in this study), and other groups (such as non-academic users of the research literature) could usefully be included in future studies. Our coding was undertaken using agreed protocols and involved a process of validation provided by three different members of the authorial team but necessarily involves interpretation and judgement on the part of the researchers.”

Semi-structured interviews. Semi-structured interviews have ‘spine’ of common questions running through the interview shared by all participants (or all participants in a particular group), but allow room for the interviewer to pursue areas of interest arising from participant responses, including probing for greater clarity, where needed. It is probably the most common interviewing approach for qualitative research. We have added a brief explanation to the paper and also a reference to a more detailed explanation should the reader be interested to follow it up:

“Interviews were conducted using a semi-structured approach – incorporating a ‘spine’ of common questions for all participants, and some questions specific to different actor groups – allowing room for the interviewer pursue areas of interest arising from participant responses, including probing for greater clarity, where needed (Bryman, 2015).”

Sampling approach. We have described purposive heterogeneous sampling above in our introductory explanation. We have expanded the explanation in the paper itself to clarify this.

“The study adopted a heterogeneous purposive sampling approach, aiming to include a wide range of perspectives from actors in the area, selected in a “strategic way” in order to address the objectives of the study (Bryman, 2015). The sample was heterogeneous in a number of respects: firstly, it contained representatives of different roles in the scholarly communication system; secondly, it included participants from different countries (and therefore policy environments); thirdly, it comprised interviewees and with different views and levels of experience of using preprints. Participants comprised senior representatives from research funders, research-conducting organisations (universities and research institutes), preprint services, other related service providers (such as infrastructure providers), as well as researchers, both researchers demonstrably engaged with preprints (they had themselves posted a preprint) and non-engaged (there was no evidence of them having posted a preprint)…” etc.

Sample size. Sample size is addressed in our introductory explanation. 38 is a relatively large number for a study of this kind, enabling the study to achieve the desired heterogeneity and theoretical saturation. We have added a sentence explaining this in the Method: “The sample allowed the study to achieve the desired heterogeneity of actors and perspectives.”

  Reviewers’ comments:

  • Please provide more information about the subjects to help readers assess relevance:
    • What is their geographical context?
    • How knowledgeable were subjects, and do they bring bias for or against open science into their statements? 
    • How did interviewees arrive at their opinions? How much close experience do they have with preprints? Were participants asked to answer only within their own expertise? 

 

Authors’ response:

All of our participants and their affiliations are named and so readers can draw judgements from this – this information is available in the extended data. As already discussed, we have an ethical commitment not to associate individuals with specific quotations, which is conventional for research of this kind as explained in the introductory explanation above.

Reviewers’ comments:

  • There are quoted statements that reflect inaccurate understandings and are not always qualified by the author: for example, “relating to whether preprints are seen as an appropriate object for evaluation in exercises such as the UK’s Research Excellence Framework, about which there remains ambiguity. One representative of a university stated: “they’re not acceptable for REF, so they’re not even part of the equation. So it’s the author’s accepted manuscript is the currency we deal with.” (Research performing organisation)”  This is inaccurate and misleading (see https://www.ref.ac.uk/publications/guidance-on-submissions-201901/) but there is no further commentary from the authors. Please add clarification that the interviewee responses are not necessarily factual. (It would be interesting to compare interviewee opinions with known facts to understand which topics are well-understood.)

 

Authors’ response:

On the specific issue of the REF and preprints, the guidance was been updated at the beginning of 2019 whilst we were still undertaking our interviews. We have added a comment in the paper clarifying the situation, but acknowledging there is still ambiguity in this area in terms of practice: the official rules of the REF and actual practice may often be in tension:

“The concept of “standing” (Neylon et al., 2017) is relevant here, with, as an example, standing relating to whether preprints are seen as an appropriate object for evaluation in exercises such as the UK’s Research Excellence Framework, about which there had apparently been uncertainty. One representative of a university stated:

“they’re not acceptable for REF, so they’re not even part of the equation. So it’s the author’s accepted manuscript is the currency we deal with.” (Research performing organisation)

The REF guidance was updated in January 2019 (whilst our interviews were still ongoing) and now states that preprints can be included in REF submissions, but expresses a preference for “final versions” of papers rather than preprints (Research England, 2019). It is, however, also possible that institutions have developed local conventions for REF submission which in fact discourage submission of preprints, as might be inferred from this quotation.”

In some ways, there is a balance to be achieved here. Qualitative research aims to find out about perceptions and this can include misunderstandings. It is important to report these, as misunderstandings often shape behaviours. The extent to which the researcher should correct the participants in their reporting is a moot point in general. That being as it may, in this case, we do recognise there was a need to comment.

Reviewers’ comments:

  • Clarity of presentation: 
    • Please provide more quantitative context to the opinions expressed. For example, for phrases such as “Some researchers” it would be useful to be more specific. Please also be more quantitative in assessment of the themes — it is useful to know if a sentiment was expressed by a majority or minority, as shown clearly in tables of benefits and challenges. Please provide aggregate data of thematic analyses, not just quotes, and use specific numbers and proportionality (Two of Five researchers mentioned X as a benefit).

Authors’ response:

We have addressed this comment by amending wording to be as precise as possible within the constraints of a qualitative study. Once again, there is a balance to struck here. On the one hand, as a general principle, it is essential to recognise that qualitative research should not be analysed quantitatively. Inferences cannot be drawn from the data in any statistically valid way from quantitative analysis of a dataset like this, and it is important not to give the impression that they can. Certainly, including percentages would not be appropriate. In any case, the data is not sufficiently structured to allow for reliable identification of a type of response which could be reported quantitatively. Every datum in a qualitative dataset is actually unique and although it is possible to identify commonalties, it is not normally possible to say that a certain proportion of participants said one thing or another. This likely to lead inappropriate homogenisation of views.

Furthermore, it is important not to confuse prevalence with significance. In a qualitative dataset, some ideas may be mentioned by a large number of participants, but this does not necessarily mean they are significant. Significance may not, of course, be easy to judge (and it is a matter of judgement based on careful analysis) but it is not a simple issue of numbers of times an issue is mentioned in the dataset. We have however made some attempts to identify prevalence of issues raised in discussing pros and cons of preprints, something which is possible (simply identifying topics mentioned), with the relevant caveats.

On the other hand, it is perfectly reasonable to mention some quantitative terms, “the majority”, or “a minority” or if an issue was only mentioned by one or was mentioned by all participants. That is what we have tried to do within the bounds of what is defensible.  

Reviewers’ comments: 

  • The paper combines elements of a literature review, white paper/policy piece, and a qualitative study. It would be helpful to clarify the origin of ideas by separating the work clearly into Results and Discussion sections. (As one example, under the heading “Financial sustainability and business models,” it is unclear whether the proposed taxonomy was expressed by interviewees or generated by the authors. Similarly, it is unclear whether use of the term “seminal” comes from interviewees or authors.)

Authors’ response:

This point is well made. The systematisation of the different models or scenarios of delivery of preprint services is ours based on a composite analysis of our data. Many of our participants talked about ownership and sustainability issues but none of them organised their remarks in this way. We had thought the systematisation of views helped to contextualise the reporting of the data but it does create the potential for confusion, as the reviewers point out. So we have, therefore, pulled this section out of the findings and added it into the discussion in order to make it clear that it is our analysis arising from the data not a direct reporting of the data.

Reviewers’ comments:

  • Readability:
    • Please use active voice and concise language to improve readability.

 

Authors’ response:

We are conscious that conventions vary in this area across disciplines or even journals. Many life sciences researchers have moved over recent years to using the active voice. However, we were not aware of there being a house style for F1000. The discursive style we have adopted using a combination of active and passive voice is conventional for the reporting of the research we have carried out. We have, however, made some adjustments to improve clarity.

Reviewers’ comments:

  • Bring interview questions and list of participants into this report — it is hard to interpret findings while having to continuously refer back to other sources, e.g. the 2019b interview structure paper.

Authors’ response:

We believe the comments from the two sets of peer reviewers are somewhat in tension with regard to the level of detail that should be included, with the other reviewers stating:

The reviewers wonder if the results section as a whole, which is relatively long in comparison to the rest of the paper, should be limited to Table 5 and its description and avoid the unending narrative of anecdotal specific responses.

Given the insertion of the interview questions and list of participants would make the paper still longer, we believe it remains appropriate to make them available as separate, citable objects which are referenced by the paper.

Finally, sharing interview questions and participants via ‘extended data’ is required by the F1000Research Author Guidelines, which state that: “F1000Research does not accept supplementary material. Additional materials that support the key claims in the paper but are not absolutely required to follow the study design and analysis of the results, e.g. questionnaires, or supporting images or tables, can be included as extended data.”

  Reviewers’ comments:

  • Claims: 
    • The abstract sets the reader up to expect a review of preprint servers that have emerged since 2013. However, the methods section discloses that coverage of the study is limited to biology, chemistry, and psychology. It would be helpful to readers to state this up front. It is also unclear how this scope was applied in the selection of study subjects. It is difficult to determine if the scope influenced the selection of participants as listed in https://zenodo.org/record/2654832#.XTeUB5NKjs0 – the questions listed at https://zenodo.org/record/3240426 do not seem specific to these three disciplines and the quoted answers in this report suggest interviewees’ responses were not limited to these three disciplines. Please clarify how the scope of the interviews was constrained to the stated subjects, or if not, please revise the stated scope to more accurately reflect the scope employed.

Authors’ response:

We have changed the text at the beginning of the methods section to provide greater clarity about our approach, as follows:

“We chose to carry out detailed interviews of key actors in this space who could explain in depth their perceptions, attitudes and practices in relation to preprints. Participants were asked about their perspectives on preprints in general, but we intentionally recruited interviewees (where they had disciplinary affiliations) particularly from disciplines where preprint services are relatively new and rapidly growing. These were biology, chemistry and psychology, corresponding to preprint servers, bioRxiv, ChemRxiv and PsyArXiv. Focusing on these areas helped us to gauge the impact that preprints are having in areas where they are more innovative, and since many of our participants were able to speak more generally about preprints, we were able to draw comparisons with disciplines where preprints are established and which are better represented in the literature (e.g. physics, computer science, and economics).”

It is important to note that our study was not designed to elicit comments about specific servers in particular but rather about participants views on preprints in general. Their own experiences, based as some of them were in particular disciplinary communities, will, of course, be coloured by their familiarity with particular servers, but we wanted to explore their views on preprints more generally than that.

Reviewers’ comments:

  • The abstract does not reflect the content. The interview responses are summarised in general terms, not related to defined scope or the newer servers. A substantial proportion of this article is review material: either in the explicit literature review section, or as author-contributed content interspersed with summaries of interviewee responses. For example, conclusions listed in the abstract appear to be drawn from the literature review.

 

Authors’ response:

We have made changes to the abstract that we hope correct this, including ensuring findings reported relate closely to the data (although for the most part these do correlate with the literature).

Reviewers’ comments:

  • “Our study is the first using empirical data to understand the new wave of preprint servers” — this is not supported by the study content: no results are specifically related to the newer servers identified in the introduction; quoted excerpts indicate that interviewees responded in general about all servers and disciplines. Furthermore, prior surveys and bibliometric analyses have looked at a subset of these servers, either individually or in groups (eg https://osf.io/5fk6c/).1

 

Authors’ response:

We have made changes to the abstract that we hope correct this.

Minor comments

Reviewers’ comments:

We recommend making the following amendments to improve the clarity and accuracy of the report.

In the Introduction and literature review: 

  • The summary of the longer history of preprint efforts across disciplines, would benefit from incorporating and citing ‘The Brief Prehistory of Preprints” (Cobb, 2017)2

Authors’ response:

Added.

Reviewers’ comments:

  • Zuckerberg Foundation should be corrected to Chan Zuckerberg Initiative.

 

Authors’ response:

Corrected.

Reviewers’ comments:

  • Discussion of new servers would be improved with a note as to their current size, and with awareness that not all new ‘preprint’ servers are exclusively for preprint content (many COS servers include postprints that are not possible to easily distinguish from preprint versions).

Authors’ response:

We have added figures for bioRxiv, ChemRxiv and PsyArXiv.

Reviewers’ comments: 

  • “This paper aims to explore…by investigating current practices, drivers and barriers to [preprint] use.” The interviews indicate perceptions and attitudes but do not reveal actual behaviours or practices. Please clarify how this study examines current practices.

 

Authors’ response:

The data covers behaviours and practices in a number of ways. Most prominently, the data indicates differences in uptake of preprints by researchers, with some being engaged and talking about their experiences and the positive outcomes that have accrued in their view. Others were sceptical and spoke about their lack of engagement based on their views of the some of the challenges presented by preprints. The thematic area covering perceived benefits and challenges has experience of adoption (or non-adoption) running through it. The theme of “Disciplines, cultures and practices” also has behaviours and practices throughout.  In it, we refer to differing levels of awareness and “adoption”, “a willingness to experiment” etc. and present illustrative quotations from interviews. The earlier section on definitions also has some elements of behaviour, with one participant, for example, talking about posting a preprint when it was “submission ready for a journal”. The behaviours and practices are then closely linked to and shaped by the perceptions and attitudes, as might be expected. We have tried to make this clear at the beginning of the sub-section on benefits and challenges: “Participants highlighted a number of (potential) benefits and challenges of preprints which were seen to relate to particular practices around adoption or non-adoption of preprints”. 

Reviewers’ comments:

Figures and tables:

  • Fig 1: To increase accessibility, please provide as raw text with headings or tags to indicate which stakeholder group you believe is most relevant

Authors’ response

Figure 1 is now available in csv format (text only) via Zenodo. This is referenced in the text as Chiarelli et al 2019c - http://doi.org/10.5281/zenodo.3539032.

Reviewers’ comments:

Literature review:

  • The authors focus on four main issues — how were these arrived at? Was the literature review systematic in any way? 

Authors’ response

We did not perform a formal systematic literature review but our reading of the literature is that these issues emerge. We have added a new category to cover the more empirical studies now emerging, mostly in 2019.

Reviewers’ comments: 

  • Given the emergent phase of preprint infrastructure in biology, chemistry and psychology, and as the authors note that empirical data is lacking, it may be equally valuable to include more grey literature (blogposts, webpages) in this review, given that several of the cited articles are opinion pieces, albeit as editorials or peer-reviewed review articles. Of note, we find personal blogs are a useful source of stories that detail benefits and drawbacks of preprints from personal experiences.

 

Authors’ response

We have included some of these more informal sources, including information form the ASAPbio web site, Scholarly Kitchen blog, SciELO blog and numerous news sections of journals. We have added an example of an exchange on Twitter. We do not claim to be exhaustive but have been intentionally selective to represent the main strands of debates. Including many more would make an already long article even longer when we would prefer the main focus of the article to on the data we present.

Reviewers’ comments:

  • "Dis-benefits:" is not a common term, perhaps use “drawbacks,” “concerns,”, “disadvantages” or “potential risks”.

Authors’ response:

We have changed the term to “challenges”.

Reviewers’ comments:

  • When discussing the lack of quality assurance, the authors seem to make an implicit assumption here that journal-led peer review assures quality. It would be fair to contextualise this section with some literature on peer review to understand variety of peer review outcomes on quality of paper and validity of findings.

Authors’ response:

We have added reference (Lee, et al., 2013) in the literature review on problems with peer review. Concerns about quality were raised by our participants, which we report, although as we also say, many were eager to report the imperfections of peer review as well.

Reviewers’ comments:

Results:

  • In the discussion of preprints in the medical field, please note that the concerns raised are being addressed by medRxiv. https://www.medrxiv.org/

Authors’ response:

medRxiv had not been released when our data was collected. Our data are inevitably a snapshot from the time it was collected. However, we note that the quotation from this area in the paper covers many of the checks now being carried out by medRxiv.

Reviewers’ comments: 

  • In the table of business models: 
    • Note that bioRxiv uses 3rd party technology, Highwire.

Authors’ response:

Corrected

Reviewers’ comments:

  • “Publisher driven” models are not new; PeerJ is older than bioRxiv, Nature had Nature Preceedings in the early 2000s, and bioRxiv has had J2B preprints since its inception. 

Authors’ response:

Noted and details about the closure of PeerJ preprints also now added.

Reviewers’ comments:

  • How does the outsourcing of infrastructure to a third party enable sustainability? Arguably, depending on the third party, this leaves the service vulnerable to changes in service entirely outside its control.

Authors’ response:

We have added clarification on this, as follows:

“Model 2 is a version of 1 in which some infrastructure is outsourced to a third party, something which might help to enable sustainability by creating efficiencies through economies of scale (a common benefits of outsourcing). The infrastructure provided by the Center for Open Science is an example of such benefit, with multiple preprint servers being run by the same organisation and on the same infrastructure, avoiding duplication, and therefore creating efficiency.”

Reviewers’ comments: 

  • Individual authors can still submit to PeerJ.

Authors’ response:

We have updated references to PeerJ following the announcement of their closure.

Reviewers’ comments: 

  • Most publisher-mediated preprint submissions are voluntary for authors (a notable exception being F1000).

Authors’ response:

We have added a note to this effect in the Introduction.

Reviewers’ comments: 

Authors’ response:

We have reviewed the survey results at the URL cited and agree they are useful and interesting. However, we cannot see substantial information about “subject preferences” in the data. We have not therefore made any changes in response since we are unsure what the recommendation of the reviewers refers to.

The following suggestions may also help improve the paper but are not essential.

Reviewers’ comments:

In the Introduction and literature review:

  • Open Research platforms could be discussed in context of F1000Research, since this is the model they are derived from (in collaboration with F1000). 

Authors’ response:

The Open Research Central services are further examples of the model developed by F1000. Since we are not attempting to be exhaustive in listing of services but are rather giving examples, and because Open Research Central uses the same model as is “powered by” F1000, we would prefer to keep our introduction at the length it is without giving more examples, which would further extend it.

Reviewers’ comments:

  • Please clarify the age of PeerJ and Preprints.org in comparison with bioRxiv.

Authors’ response:

We have done this, adding a note of the launch (and closure) of PeerJ in the opening paragraph.

Reviewers’ comments:

  • PLOS is not the only journal that deposits preprints to biorxiv using the J2B mechanism — several do, please refer to the list of journals offering direct transfers to bioRxiv (https://www.biorxiv.org/about-biorxiv).

Authors’ response:

We have done this, adding a reference in the opening paragraph.

Reviewers’ comments:

 

Reviewers’ comments:

Figures and tables:

  • Fig 1 introduces some interesting questions, however many of these are related to topics that are brought up for the first time in this figure.

Authors’ response:

Yes, ours was an exploratory study which had as one of its aims to identify areas requiring further work. Our data pointed to a wide range of questions which now need to be pursued in order to understand more about the (potential) future of preprints. What we have done is to help to provide clarity around what the they key questions are and how they might begin to be addressed (e.g. which actors should be involved in addressing them). This was designed to be helpful to policy makers in particular.

Reviewers’ comments:

Authors’ response:

Updated

Reviewers’ comments:

Literature:

  • “However, Bourne et al. (2017) controversially extend the definition to include “a paper that has been peer reviewed and…was rejected, but the authors are willing to make the content public”.” — Please provide reasoning or citation as to why this is controversial.

Authors’ response:

The context as a whole is important: “For (3) versioning, the relationship of a preprint to peer review is central. Desjardins-Proulx et al.’s (2013, p. 1) observation is typical in stating that preprints are made available, “before, or in parallel to, submitting them to journals for traditional peer review”. Suber (2012, p. 102) points out that this is not to “bypass peer review”, but that it applies to “works destined for peer review but not yet peer reviewed”. However, Bourne et al. (2017) controversially extend the definition to include “a paper that has been peer reviewed and…was rejected, but the authors are willing to make the content public”.

The point is that if preprints are defined as papers “destined for peer review” then including papers was rejected in a peer review process extends the definition and does so in a controversial way because it might imply a by-passing (or ignoring) of the peer review process. We note that in our data some participants questioned whether an output can be considered a preprint if it is not ultimately formally published.

Reviewers’ comments:

  • Accessibility — note that preprints are not necessarily open access, where they are not provided with permissive licensing, and this is certainly the case for bioRxiv.3

Authors’ response:

Point noted. We have changed the description in Table 1 to, “Preprints are openly available online”. In the associated discussion we talking about availability of content but do not use the term ‘open access’.  This was intentional, as it allows us to side step the question of definitions of open access, which is a controversial issue (gratis, libre etc). We appreciate the point about licensing and this was one issue mentioned in our dataset and to which we make reference.

Reviewers’ comments:

  • One potential benefit to ECRs that is not mentioned here is being able to demonstrate productivity to funders and hirers. This would be useful to append to the current paragraph, given it is raised in the interview questions.

Authors’ response:

Good point, and as this is a somewhat different point than achieving visibility, we have added it to the paragraph:

“Early dissemination is seen by some as especially useful to a number of members of the scholarly community in particular, with early career researchers (ECRs) commonly identified as specific potential beneficiaries, as preprints can allow them to rapidly achieve “visibility” and demonstrate productivity in job and grant applications (Desjardins-Proulx et al., 2013; Sarabipour et al., 2018; Tennant et al., 2019).”

Reviewers’ comments:

  • “There also appears to be little acknowledgement that the claim stands in tension to the one cited earlier that preprints often differ little from final published versions.” — There is only tension if both claims concern all preprints. It is possible for there to be preprints that improve through a lot of feedback as well as preprints that undergo no/very little change.

Authors’ response:

We have updated this sentence to read “There also appears to be little acknowledgement that the claim, when used to make the case for preprints in general terms, stands in tension to the one cited earlier that preprints often differ little from final published versions.”

Reviewers’ comments:

  • Please use unshared or unpublished instead of “homeless”.

Authors’ response:

“Homeless” has been used elsewhere in the context of preprints (e.g. Berlin, 2018) and more generally in the context of scholarly communication. Mega-journals have been described as “allowing for the publication of research that might not find a home in traditionally more selective journals” (Wakeling, et al. 2017). That definition helps to make the point that “homeless” does not just mean outputs that are unshared or unpublished but ones that might not otherwise find a venue for publication. We therefore propose to retain the term.

Berlin, S. (2018). If the papers don’t come to the journal…. EMBO Reports, 19(4). https://doi.org/10.15252/embr.201845911

Wakeling, S., Spezi, V., Creaser, C., Fry, J., Pinfield, S., & Willett, P. (2017). Open access megajournals: The publisher perspective (Part 1: Motivations). Learned Publishing, 30(4), 313–322. https://doi.org/10.1002/leap.1118

Reviewers’ comments:

  • We recommend avoiding the term “policy stack,” as there is no dependent relationship among these different groups/stakeholders.

Authors’ response:

We have removed the term from the paper.

Reviewers’ comments:

Methods:

  • This section ignores the efforts of researchers as change agents.

Authors’ response:

We have now acknowledged the potential for change agents to affect community norms within our description of the ‘nature of the social system’.

Reviewers’ comments:

Results:

  • Please provide evidence for this statement: “we note that adjacency between disciplines appears to be playing an important role.” 

Authors’ response:

We have adjusted the wording of this to make sure it is clearly defensible:

“Finally, we note that adjacency between disciplines may play an important role. For example, disciplines close to those which posted preprints (e.g. those close to some areas of computer science or physics, which use arXiv consistently) may be more favourably disposed to the practice compared to those from other disciplines.”

The extension of arXiv to other adjacent disciplines is evidence of our point and we see hints of this elsewhere – but at early points.

Reviewers’ comments:

  • “Many of the participants agreed that preprints servers could be a useful outlet for otherwise ‘homeless’ research outputs, even though this goes counter to the emphasis of many of preprints as early versions of outputs later formally published elsewhere.” — this attitude may reflect a lack of publication venues for such content, not a lack of desire to publish these other outputs in a peer-reviewed destination.

Authors’ response:

Yes, agreed (see our remarks on “homeless” above). It does shift the model of preprints as being early versions of papers that are later formally published elsewhere, however.

F1000Res. 2019 Jul 23. doi: 10.5256/f1000research.21512.r50461

Reviewer response for version 1

Sarvenaz Sarabipour1, Humberto Debat2

Summary:

In this manuscript Chiarelli et al. assess preprints from the point of view of a diverse group of stakeholders. The authors perform an analysis based on 38 interviews with representatives of research funders, research performing organizations, preprint servers and service providers; and researchers (engaged and unengaged). Additional discussion is provided on the benefits and challenges of preprint posting, along with issues such as infrastructure and financial sustainability and the definition of a ‘preprint’ in different communities, and the impact this has on further uptake. This study provides a thorough investigation on an emerging and relevant topic, giving a platform to the opinions of under-represented stakeholders on the discussion of preprints: a key driver of the transformation of the scholarly publishing landscape.

Comments:

We would like to ask the authors to address the following:

The introduction may benefit from additional literature to balance the many sections. For instance:

1. On Page 3: “However, skeptics have questioned the value of preprints and even suggested they may be dangerous – circulating versions of articles before they have been quality controlled by peer review may lead to unnecessary risk, particularly in disciplines like medicine (Sheldon, 2018).” The only negative note on preprints by Sheldon 2018 was well rebutted by at least 5 publications:

  • Maintaining confidence in the reporting of scientific outputs - Sarabipour et al. (2018) 1.

  • Preprints are good for science and good for the public - Sarabipour (2018) 2.

  • Preprints help journalism, not hinder it - Tennant el al. (2018) 3.

  • Together scientists and journalists can spot poor preprints - Fraser & Polka (2018) 4.

  • Preprints: recall Nature’s nasty past - Cobb (2019) 5.

Articles 1-4 above address the relationship between scientific reporting and journalism. Further to the authors remark on “unnecessary risk particularly in disciplines like medicine” - preprint servers carry a highly visible note on preprinted manuscript stating that these research products are not peer-reviewed.

Regarding biomedical and medical preprints and risk to public: This is not a compelling argument since care is delivered to patients by physicians. MedRxiv was recently established and is accepting manuscripts to accelerate medical research. The server front page notes that: "Caution: Preprints are preliminary reports of work that have not been peer-reviewed. They should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.”

2. On Page 3: “the current and potential future role of preprints as a vehicle for scholarly communication” - there are and will be many role(s) for preprints in the current and future of the scholarly endeavors and they are summarized here:

  • On the value of preprints: An early career researcher perspective - Sarabipour et al.(2019) 6

  • https://asapbio.org/reading

  • In praise of preprints - Fry et al. (2019) 7.

3. On Page 4: “much of the literature is still to be found in editorials and opinion pieces rather than data-driven research” - the fact is that the number of data driven literature on preprints is growing rapidly and are noteworthy. Examples are:

  • Tracking the popularity and outcomes of all bioRxiv preprints - Abdill & Blekhman (2019) 8.

  • Releasing a preprint is associated with more attention and citations - Fu & Hughey (2019) 9.

  • The effect of bioRxiv preprints on citations and altmetrics - Fraser et al. (2019) 10.

  • Comparing quality of reporting between preprints and peer-reviewed articles in the biomedical literature - Carneiro et al. (2019) 11.

4. On Page 4: “Accessibility (4) is crucial, with a preprint normally defined as being (or assumed to be) openly available: it “can be viewed without charge on the Web” and “Therefore, the venue for distribution of preprints is often assumed to be a freely-accessible server of some kind”. It is a fact that all preprint servers are freely accessible to any human being that has access to internet around the world: free of charge to submit by authors, free of charge to read by all anywhere on the planet.

5. On Page 5: “To say that an output is “deserving” of dissemination is, of course, a value judgement and difficult to demonstrate for each deposit as it is made, but it is one that is implicit in much of the discourse on preprints.” - this is not entirely true. The value of preprints are well known/articulated/discussed at this point as preprints accelerate research and altmetric attention to scientific work: Please see references 8-11 above. Preprints have been deposited on arXiv for nearly three decades now and are invaluable to the physical sciences community. Other forms of pre-peer review material such as computer code and protocols have been disseminated open source and their value is well noted in various fields.

   

6. On Page 5: “dis-benefits of preprints” - unless the authors may provide any evidence of “dis-benefits of preprints” besides one note (Sheldon 2018), it is best to avoid using this word “dis-benefits” and consider rephrasing to “perceived dis-benefits”. Please see reference 1-5 above.

7. On Page 5: “Early dissemination can be useful to some particular members of the scholarly community, with early career researchers (ECRs)” - preprints and all other forms of scholarly pre-publication material are beneficial to ALL researchers not jut ECRs. Timely release of scientific results in the form of preprints and other forms such as code and data bases have already accelerated research by encouraging sharing of ideas, resources and as discussed in previously suggested references above.

8. On Page 5: “Perhaps the most prominent criticism of preprints relates to this last issue: the lack of quality assurance through peer review (Sheldon, 2018). As well as a general concern about lowering quality standards, lack of quality control has been seen as potentially dangerous as “reports that have not undergone formal peer review [organised by a journal] could be misleading” (Lauer et al., 2015). Furthermore, uncertified science might be reported prematurely in the media and might even give rise to ‘fake news’ (Sheldon, 2018). Some insist that, at the very least, the opportunity to disseminate knowledge rapidly without peer review may encourage academics to produce low-quality outputs on fashionable topics (Teixeira da Silva, 2017).” The authors should declare that there are no evidences supporting all these bold statements. The Sheldon (2018) news article in fact explicitly states its speculative nature in the title which reads “preprints could promote confusion and distortion”. Please see references 1-5 above.

9. On Page 5: “Others have gone further and questioned the value of self-appointed reviewers, as opposed to those selected by journal editors (see the issue of “self-policing” highlighted by Harnad, 1998).” This is a misleading claim. Voluntary reviews of preprints are done by researchers in that field or community and are not self-appointed by the authors. The open access nature of preprints allows the potential assessment of manuscripts by more than 2-3 (more than journal reviewers) normally and there is no evidence on these reviews having different quality than journal peer-reviews.

10. On Page 5: “Preprint posting, however, is unlikely in any case to substitute for the valuable role played by selective journals in filtering content”. This is an unfounded claim as a system of preprint servers and overlay journals may very well replace the outrageously expensive and time-consuming system of for profit journal publishing in under a decade. Please see reference 12.

  • A proposal for the future of scientific publishing in the life sciences - Stern & O’Shea (2019) 12.

11. On Page 5: “Whilst this convention has come under criticism and been withdrawn by some publishers, it still exists for some journals, e.g. in medicine and chemistry (Lauer et al., 2015; Teixeira da Silva & Dobránszki, 2019)” - this sentence is misleading. In the current scenario, authors should consider changing “some” to “most” publishers.

12. On Page 6: “For example, it is not universally agreed when an output should be citable (in the literature, funding proposals or promotion cases)”. It is now becoming more and more agreed upon that preprints should be cited in articles and research grant proposals. Major funding bodies such as NIH, CZI and Wellcome Trust allow and encourage citation of preprints. Many journals allow citation of preprints and other research outputs (eLife and PLOS family of journals are examples).

 13. On Page 6 and other pages: “Scooping”, and “a claim of precedence”: Scooping is a perceived concern of researchers and is unfounded. Many researchers present unpublished research at conference. Perceived preprint concerns and claim of precedence have been discussed in reference 6 above.

14. On Page 6: “or when it can be used to establish a claim of precedence” and “In disciplines where a preprint is not considered appropriate to establish precedence, it has also been suggested that making a preprint available may actually encourage research to be scooped by rival researchers who publish in a recognized journal before the preprint authors (the ‘flip side’ of the priority claim argument above) (Kaiser, 2017).”Regarding these repeated sections on scooping and priority claims:

  • 2-3 papers published via preprints or journals in close temporal proximity of each other have all been projects 2-4 years in making so it is very difficult to assess who did it first.

  • Who gives researchers priority? A preprint that is posted online has a date and that gives precedence to the work. In reality, as researchers we all know that priority is given to researcher by their field and their peers. Scientific research is often presented at conferences well in advance of preprints or journal publication. As we know what matters in research is scientific reputation and that is given to researchers by their peers and field during a career in scientific research.

  • Multiple papers published via preprints or journals in close proximity of each other only show the reproducibility of one another are need to be celebrated not feared so repeated arguments on scooping and priority claims do not move a positive conversation on research forward specially for early career researchers and the future generations of researchers.

15. On Page 6: “Some have argued that initiatives such as Declaration on Research Assessment (DORA), with its emphasis on the quality of the output rather than venue of publication, promote use of preprints (Polka, 2018)”. It is noteworthy that DORA has now become a major international initiative and is signed by over 1,400 organizations including major universities and other research institutions and 14,000 researchers.

Other comments to the MS text:

A) The Title:  Authors could consider an alternative title “Preprints and Scholarly Communication: a preliminary/exploratory study on Adoption, Practices, Drivers and Barriers”

B) The Abstract:

  • “Our study is the first using empirical data to understand” - please change to: Our study uses empirical data… (See comments below highlighting other empirical studies on the topic).

  • “The main concerns are related to the lack of quality assurance and the ‘Ingelfinger rule’.” Please change to: main perceived concerns.

C) The Introduction:

  • On Page 3: “key actors”. It is puzzling how the authors determined which the key stakeholders are in the preprints ecosystem. Authors should discuss why they did not incorporate “scientific journal publishers” as a relevant actor for this work. It is interesting that they incorporated “publishers” as node in Figure 1 as a stakeholder group but failed to incorporate “representatives” during their interviews.

  • On Page 3: “(Carà et al., 2017) .” Please delete the space.

  • On Page 3: “The study is the first using empirical data”.  Please see our comment on the abstract.

  • On Page 3: “Zuckerberg Foundation“ - the correct name is the Chan-Zuckerberg Foundation (CZI).

  •  On Page 3: “The study is the first using empirical data”. Please see the comments on the abstract above.

  • On Page 4: “…relatively small number of peer-reviewed studies focusing on preprints – much of the literature is still to be found in editorials and opinion pieces”. The authors surely know that in general opinion pieces published in scientific journals are also peer-reviewed.

  • “Defining preprints. Key components. Table 1”. A relevant aspect of preprints that is not contemplated here is that preprints are published open access free of charge. Perhaps that important feature could be incorporated in component 4. Maybe this feature, highly relevant for unprivileged research communities, was neglected given the sampling strategy employed (see comment below).

  • On Page 8: “The study adopted a heterogeneous purposive sampling approach, aiming to include a wide range of perspectives from actors in the area. Participants comprised senior representatives…”. If the authors selected senior representatives of each stakeholder categories, one could question how heterogeneous the sampling approach was. Moreover, authors mention the use of snowball sampling, which is linked to homogeneous sampling. In addition, participants correspond to only eight OECD High-income economies according to World Bank (half of them only from UK and Germany), which again appear to be more reflective of a homogenous sampling approach. Authors should also highlight the intrinsic limitations of quality coding in terms of reproducibility.

  • On Page 8: “We undertook 38 semi-structured interviews”. Authors did not describe how sample size was determined. Authors should discuss how this small sample size might redound in uncertainty issues. The complete absence of statistical analyses of the data is striking. This affects the “empirical power” of the work, and one could question whether the described “trends” indicated in the manuscript are purely anecdotal, derived from sampling error and/or speculative.

  •  “Like many kinds…” paragraph. This important paragraph highlighting the limitations of the study should be moved to the discussion/conclusion section.

D) The Results section in general:

The reviewers wonder if the results section as a whole, which is relatively long in comparison to the rest of the paper, should be limited to Table 5 and its description and avoid the unending narrative of anecdotal specific responses. In case the authors feel the need to present all these quoted responses, perhaps they could be moved to the supplementary section.

  • “Some participants…” “Many saw a preprint as…” “Other participants saw preprints..” “there were signs this…” and elsewhere. This vague and unclear wording could be avoided by the use of absolute numbers or percentages of total responses.

  • Table 5: Please consider complementing the legend to provide definitions to “systemic” and “individual”.

  • Table 6: Following Table 5 title, it would be more reasonable to entitle Table 6 as “ potential challenges of preprints”, or “possible” as depicted in the same table in Chiarelli et al. 2019.

  • On Page 13: “One preprint service provider observed that only 10% of preprints received comments”. While this % is depicted as low in the text by the use of “only”, authors should contextualize this type of assertion. Peer reviewed articles published in journals are far less commented than preprints as reported elsewhere.

  • On Page 14:  “author might invite comments from people who are expected to be positive about their work”. Again, this and other speculative statements should be leveled when they are employed to contrast with traditional peer-reviewed journal articles. It is a common practice in an important number of journals to ask authors to provide contact information of potential reviewers, which could eventually lead to the same issues expressed in the preceding statement.

  • On Page 14: “We note…”. This relevant paragraph, highlighting the limitations and anecdotal nature of the study, should be moved to the discussion/conclusions.

E) The Discussion/Conclusion: As expressed above, some parts of the results section could be moved here as they reflect more speculative comments based on the interview responses.

F) The References:

This citation  “Sarabipour et al.: On the value of preprints: an early career researcher perspective. PeerJ Preprints (2018)” Is now peer-reviewed and hosted as “On the value of preprints: An early career researcher perspective. Sarabipour et al. PLoS biology (2019)[.

We confirm that we have read this submission and believe that we have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however we have significant reservations, as outlined above.

References

  • 1. : Maintaining confidence in the reporting of scientific outputs.2018; 10.7287/peerj.preprints.27098v1 10.7287/peerj.preprints.27098v1 [DOI] [Google Scholar]
  • 2. : Preprints are good for science and good for the public. Nature.2018;560(7720) : 10.1038/d41586-018-06054-4 10.1038/d41586-018-06054-4 [DOI] [PubMed] [Google Scholar]
  • 3. : Preprints help journalism, not hinder it. Nature.2018;560(7720) : 10.1038/d41586-018-06055-3 10.1038/d41586-018-06055-3 [DOI] [PubMed] [Google Scholar]
  • 4. : Together scientists and journalists can spot poor preprints. Nature.2018;560(7720) : 10.1038/d41586-018-06053-5 10.1038/d41586-018-06053-5 [DOI] [PubMed] [Google Scholar]
  • 5. : Preprints: recall Nature’s nasty past. Nature.2019;570(7759) : 10.1038/d41586-019-01748-9 10.1038/d41586-019-01748-9 [DOI] [PubMed] [Google Scholar]
  • 6. : On the value of preprints: An early career researcher perspective. PLoS Biol.2019;17(2) : 10.1371/journal.pbio.3000151 e3000151 10.1371/journal.pbio.3000151 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. : In praise of preprints. International Journal of Systematic and Evolutionary Microbiology.2019;69(7) : 10.1099/ijsem.0.003323 1841-1843 10.1099/ijsem.0.003323 [DOI] [PubMed] [Google Scholar]
  • 8. : Tracking the popularity and outcomes of all bioRxiv preprints. bioRxiv.2019; 10.1101/515643 10.1101/515643 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. : Releasing a preprint is associated with more attention and citations. bioRxiv.2019; 10.1101/699652 10.1101/699652 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. : The effect of bioRxiv preprints on citations and altmetrics. bioRxiv.2019; 10.1101/673665 10.1101/673665 [DOI] [Google Scholar]
  • 11. : Comparing quality of reporting between preprints and peer-reviewed articles in the biomedical literature. bioRxiv.2019; 10.1101/581892 10.1101/581892 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. : A proposal for the future of scientific publishing in the life sciences. PLoS Biol.2019;17(2) : 10.1371/journal.pbio.3000116 e3000116 10.1371/journal.pbio.3000116 [DOI] [PMC free article] [PubMed] [Google Scholar]
F1000Res. 2019 Nov 14.
Andrea Chiarelli1

Summary:

In this manuscript Chiarelli et al. assess preprints from the point of view of a diverse group of stakeholders. The authors perform an analysis based on 38 interviews with representatives of research funders, research performing organizations, preprint servers and service providers; and researchers (engaged and unengaged). Additional discussion is provided on the benefits and challenges of preprint posting, along with issues such as infrastructure and financial sustainability and the definition of a ‘preprint’ in different communities, and the impact this has on further uptake. This study provides a thorough investigation on an emerging and relevant topic, giving a platform to the opinions of under-represented stakeholders on the discussion of preprints: a key driver of the transformation of the scholarly publishing landscape.

Authors’ response:

We are grateful for the comments made by the reviewers. They have been extremely useful for us in reviewing our approach. We have made changes in response to the suggestions of the reviewers wherever possible and have explained our responses in detail below.

Reviewers’ comments:

We would like to ask the authors to address the following:

The introduction may benefit from additional literature to balance the many sections. For instance:

1. On Page 3: “However, skeptics have questioned the value of preprints and even suggested they may be dangerous – circulating versions of articles before they have been quality controlled by peer review may lead to unnecessary risk, particularly in disciplines like medicine (Sheldon, 2018).” The only negative note on preprints by Sheldon 2018 was well rebutted by at least 5 publications:

  • Maintaining confidence in the reporting of scientific outputs - Sarabipour et al. (2018)1.

  • Preprints are good for science and good for the public - Sarabipour (2018)2.

  • Preprints help journalism, not hinder it - Tennant el al. (2018)3.

  • Together scientists and journalists can spot poor preprints - Fraser & Polka (2018)4.

  • Preprints: recall Nature’s nasty past - Cobb (2019)5.

 

Authors’ response:

We are happy to comment that this point has been rebutted. Prompted by this comment, we have added a paragraph in the literature review on the fact there has been vigorous debate on this and a number of other aspects of preprints. We believe it is worthwhile explicitly acknowledging this, as follows:

“It is noticeable that the literature on the pros and cons of preprints has, like many aspects of open science, given rise to robust discussion and debate. The paper by Krumholz, Ross, & Otto (2018) cited above is itself structured as a debate, with the first two authors making the case for preprints and the third expressing concerns. Sheldon’s (2018) opinion piece in Nature arguing that preprints could have a negative impact beyond the scientific community, was met with vociferous rebuttals in the letters pages of the journal the following month (Fraser & Polka, 2018; Sarabipour, 2018; Tennant, Gatto, & Logan, 2018). On social media, such as Twitter, there have also been vigorous exchanges (e.g. Twitter, 2019).”

For completeness, we have also added additional references to sources raising concerns about the adoption of preprints within medicine (Krumholz, Ross, & Otto, 2018 and Leopold et al, 2019). The article by Leopold et al is particularly significant in that it is not just a contribution to the debate but is the (re-)statement of an editorial policy for several medical journals who follow the Ingelfinger rule, clearly stating arguments against preprints. We believe this remains a live debate, and have sought to reflect this in the revised article.

Apart from this particular point, a large number of comments made by the reviewers which we respond to in detail below relate to how we present the literature and the wider debate in our introduction and literature review. It is, therefore, worthwhile clarifying our approach in general terms.

In our literature review, we report various views on preprints – their features, their perceived advantages and disadvantages etc. We do so in an evaluative way but aim to represent different perspectives. We agree with the reviewers, balance is important. We, therefore, spend as much time discussing perceived benefits as portrayed in the literature as we do challenges – and our account is constructed in a way specifically to create balance.  

Our introduction and literature review sets the scene for our research – the concern about quality and potential misuse of results features in the literature as well as in our interview data, for example, and so discussing the first of these helps to contextualise the second. Our review of the literature is not an opinion or advocacy piece (although much of the literature we cite fit these categories). Our literature review is not designed to set out a particular position (for or against) in relation to preprints, nor would that be appropriate for a study of this kind. Rather, it aims to report the literature in a balanced way.

Krumholz, H. M., Ross, J. S., & Otto, C. M. (2018). Will research preprints improve healthcare for patients? BMJ, 362, k3628. https://doi.org/10.1136/BMJ.K3628

Leopold, S. S., Haddad, F. S., Sandell, L. J., & Swiontkowski, M. (2019). Editorial: Clinical Orthopaedics and Related Research, The Bone & Joint Journal, The Journal of Orthopaedic Research, and The Journal of Bone and Joint Surgery will not accept clinical research manuscripts previously posted to preprint servers. Clinical Orthopaedics and Related Research, 477(1), 1–4. https://doi.org/10.1097/CORR.0000000000000565

Reviewers’ comments:

Articles 1-4 above address the relationship between scientific reporting and journalism. Further to the authors remark on “unnecessary risk particularly in disciplines like medicine” - preprint servers carry a highly visible note on preprinted manuscript stating that these research products are not peer-reviewed.

Authors’ response:

The point we are making here is that criticism of preprints has occurred in the literature based on the argument that they are potentially dangerous. Once again, we are reporting the debate in the literature, we are not writing an advocacy/opinion piece ourselves. We are not saying we agree or disagree with the arguments, rather we are reporting them to be part of the current debate.

As noted above, we have added two further citations to our paper to further illustrate the point that this concern has been raised in the literature (Krumholz, Ross, & Otto, 2018 and Leopold et al, 2019). In the article by Krumholz et al (another opinion piece, this time in the form of a debate between supporters and sceptics of preprints), Otto argues, “Medical research preprints are unnecessary for scientific progress and dangerous for patient care”. Interestingly, the guidance on preprints produced by COPE ( https://publicationethics.org/files/u7140/COPE_Preprints_Mar18.pdf) acknowledges that such criticisms have been made of preprints (even if it is not an argument that it supports):

“Some critics have raised concerns that preprints may have a negative impact on the credibility and public perception towards research, since the information has not been scrutinised and validated via peer review. For example, what are the risks if a preprint with a potential impact on public health is interpreted by some as established evidence?”

The fact that preprints are labelled as such, as the reviewers point out, helps to mitigate the possible misuse of preprints, but does not eliminate that danger. In an open environment, there are possibilities that medical research could be accessed by patients, campaigners and others directly. There are well known examples of this happening and of the impact of misunderstandings being magnified by social media. Even if many of the concerns are hypothetical (as our research suggests), they cannot be simply dismissed, and need to be reported in any overview of the literature.

Reviewers’ comments:

Regarding biomedical and medical preprints and risk to public: This is not a compelling argument since care is delivered to patients by physicians. MedRxiv was recently established and is accepting manuscripts to accelerate medical research. The server front page notes that: “Caution: Preprints are preliminary reports of work that have not been peer-reviewed. They should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.”

Authors’ response:

Our point, again, is that the debate in the literature raises this concern. And it needs to be acknowledged, we believe, in any fair reporting of the current debate. We refer to our response above. We would add that open-access medical content can more easily be accessed directly by non-clinicians without expert mediation. In fact, the whole literature on the ‘expert patient’ builds on this as a positive. There is however the risk of misunderstanding. That is one point made by Otto (see above) and also by Leopold et al. Note MedRxiv had not been released when we carried out our data gathering, although one participant anticipated the kind of screening and flagging of preprints mentioned in the reviewers’ comments above.

Reviewers’ comments:

2. On Page 3: “the current and potential future role of preprints as a vehicle for scholarly communication” - there are and will be many role(s) for preprints in the current and future of the scholarly endeavors and they are summarized here:

  • On the value of preprints: An early career researcher perspective - Sarabipour et al.(2019)6

  • https://asapbio.org/reading

  • In praise of preprints - Fry et al. (2019)7.

 

Authors’ response:

Agreed. These papers make the case for preprints in the current scholarly communication system. We have cited Sabipour et al (2019). We will happily add reference to the more recent explanation of the advantages of preprints by Fry et al (2019). These papers (for the most part, at least) confine themselves to discussing the role of preprints in the current scholarly communication environment rather than discussing how they might contribute to more radical change in future. We have however added to our Discussion a paragraph with some more information in this area. This discusses the overlay idea amongst others:

“The future of preprints servers and their links with the overall scholarly communication process and infrastructure remain unclear. It is possible that the recent rise in preprint services might be reversed and that preprints go through a period of retrenchment, returning to serve the core areas traditionally associated with preprint use, such as high-energy physics. A possible alternative to such a ‘retrenchment scenario’ is what might be called the ‘patchiness scenario’, where different levels of adoption exist across different fields. Patchiness may be an ongoing situation, or may be a transition stage towards a possible ‘ubiquity scenario’. For preprints to become ubiquitous, of course, requires significant cultural and infrastructural change, some of which is indicated by the data presented here. This may partly depend on the integration between preprint services and other parts of the scholarly communication infrastructure and on related cultural norms. Closer integration may give rise the possibility of more radical change in scholarly communication, creating opportunity for developments, such as overlay journals. Overlay journals have been discussed as thought experiments since at least the turn of the 21st Century (A. Smith, 2000; J. W. T. Smith, 1999) and there have been notable experiments in this area, and some ongoing services in specialised areas do exist, such as Discrete Analysis (Ball, 2015). However, we are yet to see their widespread adoption, even though the potential remains.”

Reviewers’ comments:

3. On Page 4: “much of the literature is still to be found in editorials and opinion pieces rather than data-driven research” - the fact is that the number of data driven literature on preprints is growing rapidly and are noteworthy. Examples are:

  • Tracking the popularity and outcomes of all bioRxiv preprints - Abdill & Blekhman (2019)8.

  • Releasing a preprint is associated with more attention and citations - Fu & Hughey (2019)9.

  • The effect of bioRxiv preprints on citations and altmetrics - Fraser et al. (2019)10.

  • Comparing quality of reporting between preprints and peer-reviewed articles in the biomedical literature - Carneiro et al. (2019)11.

Authors’ response:

Agreed. All of these papers have been made available in 2019 in various forms (all of them since we gathered our data, two of them since we submitted our paper to F1000). They comprise a new empirical base for preprints studies. We have adjusted our literature review to take these recent additions to the literature into account, adding a new section on ‘The use and impact of preprints’:

“The use and impact of preprints. A noticeable recent development in the literature has been publication of a number of empirical studies on the use and impact of preprints. These include Carneiro et al.’s (2019) study which compared the quality of reporting of findings in preprints from PubMed and bioRxiv against formally-published journal articles based on a number of criteria tested through a questionnaire...” etc.

Reviewers’ comments:

4. On Page 4: “Accessibility (4) is crucial, with a preprint normally defined as being (or assumed to be) openly available: it “can be viewed without charge on the Web” and “Therefore, the venue for distribution of preprints is often assumed to be a freely-accessible server of some kind”. It is a fact that all preprint servers are freely accessible to any human being that has access to internet around the world: free of charge to submit by authors, free of charge to read by all anywhere on the planet.

Authors’ response:

Agreed. And that is why we have included open accessibility in our definition of what constitutes a preprint. When discussing preprints that is the way many authors of the literature define them, or at least that is a characteristic they assume. Sometimes this is taken for granted amongst commentators but is always at least implicit. We have made changes to our discussion of definitions to make this clearer. In Table 1, we have changed the description of this component to, “Preprints are openly available online”. In the accompanying textual discussion, it now reads as:

“Accessibility (4) is crucial in definitions. A preprint is normally defined as being (or assumed to be) openly available online: it “can be viewed without charge on the Web” (Berg et al., 2016, p. 899). The idea of openness is fundamental to discussions on preprints. The venue for distribution of preprints is often assumed to be a freely-accessible server of some kind, a point highlighted by Berg et al. (2016, p. 899), who include in their definition that a preprint “is uploaded by the authors to a public server”.”

Reviewers’ comments:

5. On Page 5: “To say that an output is “deserving” of dissemination is, of course, a value judgement and difficult to demonstrate for each deposit as it is made, but it is one that is implicit in much of the discourse on preprints.” - this is not entirely true. The value of preprints are well known/articulated/discussed at this point as preprints accelerate research and altmetric attention to scientific work: Please see references 8-11 above. Preprints have been deposited on arXiv for nearly three decades now and are invaluable to the physical sciences community. Other forms of pre-peer review material such as computer code and protocols have been disseminated open source and their value is well noted in various fields.

Authors’ response:

What we are discussing here is how preprints are defined in the literature. Bourne et al. (2017) extend the definition of a preprint beyond that of a version of a peer-reviewed paper before it is peer-reviewed and formally published to include “a research output that has not completed a typical publication pipeline but is of value to the community and deserving of being easily discovered and accessed”. What we are pointing out is that it is interesting that this definition defines a preprint as something that is “deserving” of dissemination. To state that something is “deserving” of dissemination is a value judgement. To incorporate a value judgement in a definition is unusual. Here we are not discussing whether preprints are valuable or not, rather we are discussing whether value should be part of the definition of what constitutes a preprint. Of course, particular preprints may or may not be valuable. Preprints as a whole may or may not be valuable. But to define preprints as what is valuable is somewhat curious. If an output has “not completed a typical publication pipeline” but is not deemed to be valuable, does that mean it is not a preprint? And if so, who decides what is a preprint or not, and on what basis? By extending the definition in this way Bourne et al have at least created an ambiguity.

Reviewers’ comments:   

6. On Page 5: “dis-benefits of preprints” - unless the authors may provide any evidence of “dis-benefits of preprints” besides one note (Sheldon 2018), it is best to avoid using this word “dis-benefits” and consider rephrasing to “perceived dis-benefits”. Please see reference 1-5 above.

Authors’ response:

We have changes “dis-benefits” to “challenges”. We now refer to “Perceived benefits and challenges” of preprints. This is a fairer way to refer to the current debate, especially as most of the outputs we refer to discussing benefits and challenges are opinion pieces which present arguments – some more cogent than others – but only limited empirical evidence.

Reviewers’ comments:

7. On Page 5: “Early dissemination can be useful to some particular members of the scholarly community, with early career researchers (ECRs)” - preprints and all other forms of scholarly pre-publication material are beneficial to ALL researchers not jut ECRs. Timely release of scientific results in the form of preprints and other forms such as code and data bases have already accelerated research by encouraging sharing of ideas, resources and as discussed in previously suggested references above.

Authors’ response:

Agreed. We have rephrased this to make (we hope) the point we are making clearer i.e. some of the literature emphasises the particular benefits to ECRs. This is not to say preprints are not useful for other groups, it is just that the benefits for this particular group are emphasised in some of the literature. The rewording is:

“Early dissemination is seen by some as especially useful to a number of members of the scholarly community in particular, with early career researchers (ECRs) commonly identified as specific potential beneficiaries, as preprints can allow them to rapidly achieve “visibility” and demonstrate productivity in job and grant applications (Desjardins-Proulx et al., 2013; Sarabipour et al., 2018; Tennant et al., 2019).”

Reviewers’ comments:

8. On Page 5: “Perhaps the most prominent criticism of preprints relates to this last issue: the lack of quality assurance through peer review (Sheldon, 2018). As well as a general concern about lowering quality standards, lack of quality control has been seen as potentially dangerous as “reports that have not undergone formal peer review [organised by a journal] could be misleading” (Lauer et al., 2015). Furthermore, uncertified science might be reported prematurely in the media and might even give rise to ‘fake news’ (Sheldon, 2018). Some insist that, at the very least, the opportunity to disseminate knowledge rapidly without peer review may encourage academics to produce low-quality outputs on fashionable topics (Teixeira da Silva, 2017).” The authors should declare that there are no evidences supporting all these bold statements. The Sheldon (2018) news article in fact explicitly states its speculative nature in the title which reads “preprints could promote confusion and distortion”. Please see references 1-5 above.

Authors’ response:

As stated above, we have added two other references to articles critiquing preprints to this paragraph – one of these (Otto) uses the term “dangerous” in relation to preprints. The other (Leopold et al) states directly, “We believe that the benefits proposed by advocates of medical preprint servers can be better achieved in other ways, and that medical preprint servers pose serious health and safety dangers to the patients for whom are supposed to be caring”. This is a view evident in the literature (as well as widely held amongst researchers and practitioners). We, therefore, need to report it.

We refer to our comments above, that we are not writing an opinion piece here – rebutting arguments point by point we may disagree with; rather, we are reporting the contours of the debate in the literature. It is appropriate, we believe, to report this as an important argument which occurs in the literature.

Reviewers’ comments:

9. On Page 5: “Others have gone further and questioned the value of self-appointed reviewers, as opposed to those selected by journal editors (see the issue of “self-policing” highlighted by Harnad, 1998).” This is a misleading claim. Voluntary reviews of preprints are done by researchers in that field or community and are not self-appointed by the authors. The open access nature of preprints allows the potential assessment of manuscripts by more than 2-3 (more than journal reviewers) normally and there is no evidence on these reviews having different quality than journal peer-reviews.

Authors’ responses:

The phrase “self-appointed reviewers” means reviewers who are self-appointed i.e. they appoint themselves, not that the authors appoint them. We hope this clarifies our point.

Reviewers’ comments:

10. On Page 5: “Preprint posting, however, is unlikely in any case to substitute for the valuable role played by selective journals in filtering content”. This is an unfounded claim as a system of preprint servers and overlay journals may very well replace the outrageously expensive and time-consuming system of for profit journal publishing in under a decade. Please see reference 12.

  • A proposal for the future of scientific publishing in the life sciences - Stern & O’Shea (2019)12.

Authors’ response:

Proposals of overlay journals have been around for at least 20 years (A. Smith, 2000; J. W. T. Smith, 1999). One of us wrote about such possibilities in 2004 (Pinfield, 2004); ideas further developed a decade ago (Pinfield, 2009). We agree, that the overlay model has real potential but are yet to see significant adoption of it (although interesting experiments and even some specialised services do exist). However, even in an overlay model filtering for quality is important and involves peer review. In the short/medium-term this function is normally seen as coming from journals since other peer-review providers are not commonly available.

Note again, however, here we are still providing commentary on literature and are not advocating any particular approach ourselves. We have, however, changed the text of this sentence to make that clearer:

“Preprint posting, however, is not normally seen as a substitute for peer review, currently managed by journals, in filtering content (Suber, 2012), a process that is commonly valued, even if recognised to be imperfect (Lee, et al., 2013).”

Also, we have added a paragraph (as previously mentioned) in the Discussion on possible futures of preprints servers and their connections to the wider scholarly communication infrastructure, including overlay journals, since we agree, there is considerable potential in the overlay model.

“The future of preprints servers and their links with the overall scholarly communication process and infrastructure remain unclear… Closer integration may give rise the possibility of more radical change in scholarly communication, creating opportunity for developments, such as overlay journals. Overlay journals have been discussed as thought experiments since at least the turn of the 21st Century (A. Smith, 2000; J. W. T. Smith, 1999) and there have been notable experiments in this area, and some ongoing services in specialised areas do exist, such as Discrete Analysis (Ball, 2015). However, we are yet to see their widespread adoption, even though the potential remains.”

Pinfield, S. (2004). Self-archiving publications. In G. Gorman & F. Rowland (Eds.), Scholarly Publishing in and Electronic Era: International Yearbook of Library and Information Management 2004-2005 (pp. 118–145). Retrieved from http://eprints.nottingham.ac.uk/142

Pinfield, S. (2009). Journals and repositories: An evolving relationship? Learned Publishing, 22(3), 165–175. https://doi.org/10.1087/2009302

Smith, A. (2000). The journal as an overlay on preprint databases. Learned Publishing, 13(1), 43–48. https://doi.org/10.1087/09531510050145542

Smith, J. W. T. (1999). The deconstructed journal – a new model for academic publishing. Learned Publishing, 12(2), 79–91.

Reviewers’ comments:

11. On Page 5: “Whilst this convention has come under criticism and been withdrawn by some publishers, it still exists for some journals, e.g. in medicine and chemistry (Lauer et al., 2015; Teixeira da Silva & Dobránszki, 2019)” - this sentence is misleading. In the current scenario, authors should consider changing “some” to “most” publishers.

Authors’ response:

Amended.

Reviewers’ comments:

12. On Page 6: “For example, it is not universally agreed when an output should be citable (in the literature, funding proposals or promotion cases)”. It is now becoming more and more agreed upon that preprints should be cited in articles and research grant proposals. Major funding bodies such as NIH, CZI and Wellcome Trust allow and encourage citation of preprints. Many journals allow citation of preprints and other research outputs (eLife and PLOS family of journals are examples).

Authors’ response:

We agree that it is becoming more common to accept citation of preprints in articles and grant proposals, and we have cited instances of this in our paper: “Perhaps the most noticeable shift recently in terms of policy is that of funder policies. Some funders have now explicitly signalled support for use of preprints, including allowing citation of preprints in funding bids, and support their inclusion in cases for academic advancement (Berg et al., 2016; Bourne et al., 2017).” However, our statement that this is not “universally” the case is still correct and is a measured way of making the point.

Reviewers’ comments:

13. On Page 6 and other pages: “Scooping”, and “a claim of precedence”: Scooping is a perceived concern of researchers and is unfounded. Many researchers present unpublished research at conference. Perceived preprint concerns and claim of precedence have been discussed in reference 6 above.

Authors’ response:

Yes, we agree that many of the concerns expressed in the area of scooping are unfounded. However, the point is that concerns are expressed in this area, and that is what we are reporting. Like many of the other areas we report from the literature, this area is also raised in our dataset. This issue is addressed in the ASAPbio site, which also discusses the EMBO “scoop protection” initiative: “If an author submits a manuscript within 4 months of posting a preprint, EMBO will consider the work novel even if a competitor publishes similar work during that time.” The fact that such protections are put in place in this way demonstrates that the issue is still a concern for many. As an advocate of preprints, Berlin (2018) still states, “While some journals already guarantee a “scooping protection”, this is not common practice and most authors are still at risk of being “scooped””. What we are reporting here is the concern.

Berlin, S. (2018). If the papers don’t come to the journal…. EMBO Reports, 19(4). https://doi.org/10.15252/embr.201845911

Reviewers’ comments:

14. On Page 6: “or when it can be used to establish a claim of precedence” and “In disciplines where a preprint is not considered appropriate to establish precedence, it has also been suggested that making a preprint available may actually encourage research to be scooped by rival researchers who publish in a recognized journal before the preprint authors (the ‘flip side’ of the priority claim argument above) (Kaiser, 2017).” Regarding these repeated sections on scooping and priority claims:

  • 2-3 papers published via preprints or journals in close temporal proximity of each other have all been projects 2-4 years in making so it is very difficult to assess who did it first.

     

  • Who gives researchers priority? A preprint that is posted online has a date and that gives precedence to the work. In reality, as researchers we all know that priority is given to researcher by their field and their peers. Scientific research is often presented at conferences well in advance of preprints or journal publication. As we know what matters in research is scientific reputation and that is given to researchers by their peers and field during a career in scientific research.

     

  • Multiple papers published via preprints or journals in close proximity of each other only show the reproducibility of one another are need to be celebrated not feared so repeated arguments on scooping and priority claims do not move a positive conversation on research forward specially for early career researchers and the future generations of researchers.

 

Authors’ response:

Yes, these are useful and interesting points. It is not clear from them, however, how we might change our paper in response. The key point we are making in our literature review is that scooping is identified as a concern in the literature. It is also expressed as a concern within our interviews, demonstrating it is a live issue in this space. The fact that preprints are often argued to be a way of asserting priority is itself evidence that researchers fear being scooped – in that case using that fear as an argument to deposit preprints. Either way, the key point we are making in our literature review is that scooping is raised as an issue, and that is what we are reporting.

Reviewers’ comments:

15. On Page 6: “Some have argued that initiatives such as Declaration on Research Assessment (DORA), with its emphasis on the quality of the output rather than venue of publication, promote use of preprints (Polka, 2018)”. It is noteworthy that DORA has now become a major international initiative and is signed by over 1,400 organizations including major universities and other research institutions and 14,000 researchers.

Authors’ response:

Yes, once again we agree. We expect that most readers of this paper will be familiar with DORA and as it is not the focus of our study, we do not discuss it in detail. If readers wish to know more, we have now provided a reference in our paper to online information about the initiative.

Other comments to the MS text:

Reviewers’ comments:

A) The Title:  Authors could consider an alternative title “Preprints and Scholarly Communication: a preliminary/exploratory study on Adoption, Practices, Drivers and Barriers”

Authors’ response:

We have changed the title to, “Preprints and Scholarly Communication: An Exploratory Qualitative Study of Adoption, Practices, Drivers and Barriers”.

Reviewers’ comments:

B) The Abstract:

  • “Our study is the first using empirical data to understand” - please change to: Our study  uses empirical data… (See comments below highlighting other empirical studies on the topic).

Authors’ response:

Noted and changes made as outlined above and below.

Reviewers’ comments:

  • “The main concerns are related to the lack of quality assurance and the ‘Ingelfinger rule’.” Please change to: main  perceived concerns.

Authors’ response:

We have added “perceived” to our discussion of benefits and challenges of preprints (as mentioned above). Here we refer to “concerns”. Concerns, by definition, relate to people's perceptions of the challenges. “Concerns” is not being used here as a synonym of “disadvantages” or “challenges” but rather to represent perceived disadvantages. “Perceived concerns” is tautological, and so we have limited our use of “perceived” in relation to benefits and challenges.

Reviewers’ comments:

C) The Introduction:

  • On Page 3: “key actors”. It is puzzling how the authors determined which the key stakeholders are in the preprints ecosystem. Authors should discuss why they did not incorporate “scientific journal publishers” as a relevant actor for this work. It is interesting that they incorporated “publishers” as node in Figure 1 as a stakeholder group but failed to incorporate “representatives” during their interviews.

     

Authors’ response:

In our study, we did acknowledge the important role of academic publishers but did not engage them directly. Our reason for this was that the publishing community is already discussing preprints in a structured way, e.g. via the work of COPE. Furthermore, we engaged directly a representative of MDPI (to discuss preprints.org) and representatives of F1000Research (to discuss their approach to pre-publication sharing), who are running preprint services.

We have added an explanation of this in our Methodology and there is a note in the new limitations statement on this as well.

Reviewers’ comments:

  • On Page 3: “(Carà et al., 2017) .” Please delete the space.

Authors’ response:

Corrected.

Reviewers’ comments:

  • On Page 3: “The study is the first using empirical data”.  Please see our comment on the abstract.

Authors’ response:

We have removed this sentence. We believe that our study is the first qualitative study on the new wave of preprints, nevertheless. As acknowledged above several important quantitative studies have appeared (most still in preprint form) in 2019.

Reviewers’ comments:

  • On Page 3: “Zuckerberg Foundation” - the correct name is the Chan-Zuckerberg Foundation (CZI).

Authors’ response:

Corrected.

Reviewers’ comments:

  •  On Page 3: “The study is the first using empirical data”. Please see the comments on the abstract above.

Authors’ response:

Corrected.

Reviewers’ comments:

  • On Page 4: “…relatively small number of peer-reviewed studies focusing on preprints – much of the literature is still to be found in editorials and opinion pieces”. The authors surely know that in general opinion pieces published in scientific journals are also peer-reviewed.

Authors’ response:

We have changed this sentence to avoid ambiguity: “Nevertheless, much of the literature is still to be found in editorials and opinion pieces rather than data-driven research.”

Reviewers’ comments:

  • “Defining preprints. Key components. Table 1”. A relevant aspect of preprints that is not contemplated here is that preprints are published open access free of charge. Perhaps that important feature could be incorporated in component 4. Maybe this feature, highly relevant for unprivileged research communities, was neglected given the sampling strategy employed (see comment below).

Authors’ response:

This the core of key component 4. We have rephrased the description in Table 1 to make this clearer: “Preprints are openly available online” (as also described above).

Reviewers’ comments:

  • On Page 8: “The study adopted a  heterogeneous purposive sampling approach, aiming to include a wide range of perspectives from actors in the area. Participants comprised  senior representatives…”. If the authors selected senior representatives of each stakeholder categories, one could question how heterogeneous the sampling approach was. Moreover, authors mention the use of snowball sampling, which is linked to homogeneous sampling. In addition, participants correspond to only eight OECD High-income economies according to World Bank (half of them only from UK and Germany), which again appear to be more reflective of a homogenous sampling approach. Authors should also highlight the intrinsic limitations of quality coding in terms of reproducibility.

 

Authors’ response:

We respond to the specific point raised by the reviewers below, but before doing so we thought it would be useful to precede our responses with an explanation of our approach to qualitative research in general. We repeat this explanation from our response to the other set of peer review comments received and although lengthy it, we hope, helps to explain our approach. This is the case since some of the comments made by the reviewers in their report are as much about qualitative research in general as they are about our particular study, and so we would like to set out our approach in the round, as context to our detailed comments further below. The open peer review process may be helpful here, as the comments we make are openly available and help to explain our approach, even though not all of the detail can be incorporated in our final paper. However, we thought it would be useful to reproduce it here as it helps to contextualise our specific responses to issues raised by reviewers below.

Our general approach in responding to the comments of the reviewers on qualitative methods is to try to achieve a balance between, on the one hand, providing clarity and further detail where necessary, and, on the other hand, avoiding over-extending an already long paper. So we have made some changes to the text of the paper with the aim of achieving greater clarity but provide more detailed explanations here in in this report in order to provide fuller explanations that might be too detailed for the paper itself.

Our study uses well-established qualitative research methods. Bazeley (2013) states: "researchers engaging in a qualitative study focus on observing, describing, interpreting, and analysing the way that people experience, act on, or think about themselves and the world around them”. Qualitative research aims to collect and analyse “rich, deep data” (Bryman, 2015) which enable understandings of issues like motivations, beliefs, and values of participants in the contexts in which they act. It focuses on understanding the perspectives of participants, including where disagreement and conflict may exist. Qualitative research typically attempts not just to identify behaviours but also meanings associated with behaviours. It is often useful for encountering unexpected issues and influences. Qualitative methods are commonly deployed therefore in studying phenomena and contexts which are emerging, where issues need to be mapped out, and theory developed (which can then in turn potentially be tested in various ways, including quantitatively). We chose to explore the preprints space using these methods in order to gain a rich and deep understanding of the issues involved, specifically including varying motivations, differing behaviours, and conflicting perspectives.

Qualitative data can take different forms and be gathered in different ways, but often take the form of text, e.g. transcripts of interviews, as in our study. In this case the transcripts are the data being analysed, and there are well-established methods for carrying out such analysis. Analysis is typically conducted inductively, identifying patterns which emerge from the data. This is normally done through a process like thematic analysis (used in this study) involving a number of steps, as explained by Braun & Clarke (2006), who are cited in our study. There are a large number of other texts that describe this process, albeit with variations e.g. Bazeley (2013), Bryman (2015), Maxwell (2013) etc. In the first stage of the process, the interviews are fully transcribed and then read and reread thoroughly by the researchers. Secondly, the entire dataset is coded – key features of the data being labelled at a detailed level. “Codes…serve as shorthand devices to label, separate, compile and organize data” (Charmaz, 1983). This is an iterative process involving constant comparisons across the dataset. It can be carried out in various ways, but in our case, we used the NVivo software which facilitates coding of large amounts of text. Coding is a kind of fracturing process, breaking down the data into small fragments, and so the third stage of thematic analysis involves assembling codes into groups or themes. Themes are identified iteratively through constant comparisons, often involved in mapping out themes in relation to one another in order to define and refine them. These were the processes we followed in this study.

In reporting qualitative research, the themes are normally used as a framework, rather than the questions that were initially asked of participants. This is an important aspect of the reporting of an inductive study in order to represent the themes that emerged from the data rather than the initial framework of the questions asked. From analysis of these themes, theoretical insights should emerge. These could take the form of systematic explanations of a context; models of relationships, issues, challenges etc; sets of questions or hypotheses that need to be further investigated; etc. We have tried to provide some of these, albeit tentatively, in our study.

Sampling in qualitative research is normally carried out in a purposive (or purposeful) way. This is described by Bryman (2015):

“Purposive sampling is a non-probability form sampling. The researcher does not seek to sample research participants on a random basis. The goal of purposive sampling is to sample cases/participants in a strategic way, so that the sample are relevant to the research questions that are being posed. Very often, the researcher will want to sample in order to ensure that there is a good deal of variety in the resulting sample, so that sample members differ from each other in terms key characteristics relevant to the research question.”

We describe our sampling approach as a “heterogeneous purposive sampling approach”, which we go on to define as, “aiming to include a wide range of perspectives from actors in the area”. The sample was heterogeneous in a number of respects: it contained representatives of different roles in the scholarly communication system, from different countries (and therefore policy environments), and with different views and levels of experience of using preprints. Participants were selected by us based on an analysis of the field, consultation with our funders and others, and then approached directly and invited to participate. We also advertised on email discussion lists and social media inviting participants, but only selected people to be interviewed who met our sampling criteria. One challenge was in finding researchers willing to be interviewed who were not engaged with preprints, and so in order to do this, we not only used the selection method of approaching people directly but also used “snowball sampling” – asking for recommendations of other people who we might approach from participants. As Bryman (2015) comments:

“Purposive sampling often involves more than one of the approaches outlined… For example, it is quite common for snowball sampling to be preceded by another form of purposive sampling. In effect, the process entails sampling initial participants without using the snowball approach and then using these initial contacts to broaden out to a snowballing method.”

This is what we did, resulting in a final sample of 38.

Sample size is a controversial issue. There is no obvious right answer to the question of, “how many interviews are enough?” The essays addressing that question by well-known qualitative researchers edited by Baker & Edwards (2012) include a wide range of different views, depending on a number of factors. Bryman quotes Mason (2010) as having examined UK and Irish PhD theses using qualitative interviews, which identifies them as using a mean sample size of 31 and a median of 28. Bryman, in his chapter on reporting research, looks in detail at a sample journal article reporting semi-structured interviews which has a sample size of 20 (Jones, Leontowitsch, & Higgs, 2010). Our sample size of 38 is relatively large. One key point to emphasise, however, is that sample size in qualitative research is normally not determined in advance. Rather, the sampling approach will be decided and participants recruited within the sampling frame. The researchers will then make a number of judgements during the data gathering process about when sufficient data has been collected. Chief amongst these is the idea of “theoretical saturation”, where new data “no longer suggests new insights” (Bryman, 2015). This involves a number of judgements, involving a close engagement with the data in relation to the research objectives.

Because qualitative research usually involves identifying areas of motivations, perspectives etc, there is commonly a need to allow participant anonymity. Often interviewees are less frank when they are ‘on the record’. Less frank interviewees tends to lead to a rather bland dataset which makes it difficult to understand the real contours of a landscape. In our study, we gained ethical approval to name participants and their organisations (with their permission) but not to associate any particular quotations with individuals. We have adopted a conventional way of dealing with this by identifying various groups amongst our sample which we do relate with each quotation (“Unengaged researcher”, “Preprint service provider” etc). Naming the group from which the interviewee being quoted came helps to contextualise each quotation, but does not compromise promised anonymity.

Providing extensive extracts of participant quotations is a critical part of presenting findings from qualitative research. Since these quotations are illustrative of the themes identified, and are examples of the data collected, showing them in detail is an important way of demonstrating the validity of the inferences being drawn from the research.

Baker, S. E., & Edwards, R. (2012). How many qualitative interviews is enough? Retrieved from http://eprints.ncrm.ac.uk/2273/4/how_many_interviews.pdf

Bazeley, P. (2013). Qualitative data analysis: Practical strategies. London: Sage.

Bryman, A. (2015). Social research methods (5th ed.). Oxford: Oxford University Press.

Jones, I. R., Leontowitsch, M., & Higgs, P. (2010). The experience of retirement in second modernity. Sociology, 44(1), 103–120. https://doi.org/10.1177/0038038509351610

Maxwell, J. (2013). Designing a qualitative study. In L. Bickman & D. J. Rog (Eds.), The SAGE handbook of applied social research methods (pp. 214–253). https://doi.org/10.4135/9781483348858.n7

To reply to the specific issue on sampling and its heterogeneity raised by the reviewers on heterogeneity and snowball sampling.

Our sample was heterogeneous in various ways and we have explained this in text added to our methodology, as follows:

“The study adopted a heterogeneous purposive sampling approach, aiming to include a wide range of perspectives from actors in the area, selected in a “strategic way” in order to address the objectives of the study (Bryman, 2015). The sample was heterogeneous in a number of respects: firstly, it contained representatives of different roles in the scholarly communication system; secondly, it included participants from different countries (and therefore policy environments); thirdly, it comprised interviewees and with different views and levels of experience of using preprints.”

On snowball sampling; snowball sampling is not linked to either homogenous or heterogeneous sampling in itself. It depends how it is used. However, as we state above, it was used in conjunction with another sampling method, as a way of achieving heterogeneity in the sample, particularly to recruit researchers (and particularly unengaged researchers).

Participants from eight different countries is a wide sampling frame for a study of this type. This study does not make any claims to cover issues to do with the Global South but does encompass a wide range of different policy environments present in the different countries covered.

We have added more explanation of the limitations of the study in a new section in the methodology, explaining limitations, as previously mentioned.

Reviewers’ comments:

  • On Page 8: “We undertook 38 semi-structured interviews”. Authors did not describe how sample size was determined. Authors should discuss how this small sample size might redound in uncertainty issues. The complete absence of statistical analyses of the data is striking. This affects the “empirical power” of the work, and one could question whether the described “trends” indicated in the manuscript are purely anecdotal, derived from sampling error and/or speculative.

Authors’ response:

We have described issues of sample size in our general explanation above.

The comments of the authors about the absence of statistical analysis from our reporting of findings needs comment. Ours is a qualitative study, using qualitative methods and qualitative analysis. It would be inappropriate to apply most quantitative approaches to our data. The “power” of qualitative findings of this sort (to use the term used by the reviewers) is in the richness of the data associated with issues of motivation, attitudes etc, not in any quantitative inferences drawn. We hope our detailed explanation of our approach to using qualitative methods is helpful in contributing to our response on this point. It includes the approach we took to inductively analysing themes in our data.

Reviewers’ comments:

  •  ”Like many kinds…” paragraph. This important paragraph highlighting the limitations of the study should be moved to the discussion/conclusion section.

Authors’ response:

Conventions on this differ across journals, and to some extent across disciplines. We have noticed statements of this sort in the method sections of other articles in F1000, and so would prefer to keep this section (in its newly amended form) in the Methods section, since it relates directly to the methods just described.

Reviewers’ comments:

D) The Results section in general:

The reviewers wonder if the results section as a whole, which is relatively long in comparison to the rest of the paper, should be limited to Table 5 and its description and avoid the unending narrative of anecdotal specific responses. In case the authors feel the need to present all these quoted responses, perhaps they could be moved to the supplementary section.

Authors’ response:

The specific responses of our participants are the data for this study. Presenting them in this way is conventional for this sort of research and doing so in detail demonstrates the validity of the inferences we are drawing. The specificity of the quotations is in the nature of qualitative research – that is where the richness of the data comes from.

It is noticeable that in his book on social sciences research, Bryman presents a sample article in order to illustrate presentation of qualitative research (Jones, et al., 2010)). This sample article follows a similar pattern of presentation of thematic findings with commentary by the authors and quotations from participant interviews. Bryman comments, “it is striking that, in presenting their findings, Jones et al. (2010) use verbatim quotations to reinforce the point they are making… They do so by including the quotations as they go along to reinforce or illustrate points they are making about the themes they extracted from that data. This is quite a common approach to the use of verbatim interview quotations.”

Jones, I. R., Leontowitsch, M., & Higgs, P. 2010. The experience of retirement in second modernity. Sociology, 44(1), 103–120. https://doi.org/10.1177/0038038509351610

Reviewers’ comments:

  • “Some participants…” “Many saw a preprint as…” “Other participants saw preprints..” “there were signs this…” and elsewhere. This vague and unclear wording could be avoided by the use of absolute numbers or percentages of total responses.

Authors’ response:

We have addressed this comment by amending wording to be as precise as possible within the constraints of a qualitative study. Once again, there is a balance to struck here. On the one hand, as a general principle, it is essential to recognise that qualitative research should not be analysed quantitatively. Inferences cannot be drawn from the data in any statistically valid way from quantitative analysis of a dataset like this, and it is important not to give the impression that they can. Certainly, including percentages would not be appropriate. In any case, the data is not sufficiently structured to allow for reliable identification of a type of response which could be reported quantitatively. Every datum in a qualitative dataset is actually unique and although it is possible to identify commonalties, it is not normally possible to say that a certain proportion of participants said one thing or another. This likely to lead inappropriate homogenisation of views.

Furthermore, it is important not to confuse prevalence with significance. In a qualitative dataset, some ideas may be mentioned by a large number of participants, but this does not necessarily mean they are significant. Significance may not, of course, be easy to judge (and it is a matter of judgement based on careful analysis) but it is not a simple issue of numbers of times an issue is mentioned in the dataset. We have however made some attempts to identify prevalence of issues raised in discussing pros and cons of preprints, something which is possible (simply identifying topics mentioned), with the relevant caveats.

On the other hand, it is perfectly reasonable to mention some quantitative terms, “the majority”, or “a minority” or if an issue was only mentioned by one or was mentioned by all participants. That is what we have tried to do within the bounds of what is defensible. 

Reviewers’ comments:

  • Table 5: Please consider complementing the legend to provide definitions to “systemic” and “individual”.

Authors’ response:

We have added definitions as follows: ““Systemic” significance relates to those factors with system-wide impact e.g. the broad scholarly communication system or disciplinary community; “individual” relates to those factors primarily affecting individuals or small groups.”

Reviewers’ comments:

  • Table 6: Following Table 5 title, it would be more reasonable to entitle Table 6 as “ potential challenges of preprints”, or “possible” as depicted in the same table in Chiarelli et al. 2019.

Authors’ response:

Agreed and done.

Reviewers’ comments:

  • On Page 13: “One preprint service provider observed that  only 10% of preprints received comments”. While this % is depicted as low in the text by the use of “only”, authors should contextualize this type of assertion. Peer reviewed articles published in journals are far less commented than preprints as reported elsewhere.

     

Authors’ response:

The preprint service provider making the statement was making the point that it was relatively low. Interestingly, it is, however, comparable to levels identified in a recent study of PLOS journals, co-authored by one of us:

Wakeling, S., Willett, P., Creaser, C., Fry, J., Pinfield, S., Spezi, V., … Medina Perea, I. (2019). ‘No comment’? A study of commenting on PLOS articles. Journal of Information Science, (Online first). https://doi.org/10.1177/0165551518819965

Reviewers’ comments:

  • On Page 14:  “author might invite comments from people who are expected to be positive about their work”. Again, this and other speculative statements should be leveled when they are employed to contrast with traditional peer-reviewed journal articles. It is a common practice in an important number of journals to ask authors to provide contact information of potential reviewers, which could eventually lead to the same issues expressed in the preceding statement.

Authors’ response:

Here we are reporting the views of participants. We have added a comment in our discussion about there being evidence of misunderstandings amongst some researchers. However, stepping in to engage with participants during the reporting of findings in this one particular instance would not be consistent with our approach or appropriate for this kind of research.

Reviewers’ comments:

  • On Page 14: “We note…”. This relevant paragraph, highlighting the limitations and anecdotal nature of the study, should be moved to the discussion/conclusions.

Authors’ response:

We have introduced a new section spelling out the limitations of our research. We again note that this is a qualitative piece of research, based on interviews, with text as the data being analysed.

Reviewers’ comments:

E) The Discussion/Conclusion: As expressed above, some parts of the results section could be moved here as they reflect more speculative comments based on the interview responses.

Authors’ response:

We have moved a whole subsection of the results into the Discussion section. This is the section which models different preprint services based on ownership and contribution.

“At a systemic level, the issue of sustainability emerged as critical from our work. Four models for delivering preprint services emerged from the data… This is a fundamental shift which has major implications for the way the role of preprints is understood and the way preprints services are configured.”

This discussion is derived from our data, but does not directly report data, but rather systematises what is expressed by our participants. It is therefore appropriately located in the discussion.

Reviewers’ comments:

F) The References:

This citation  ”Sarabipour et al.: On the value of preprints: an early career researcher perspective. PeerJ Preprints (2018)” Is now peer-reviewed and hosted as “On the value of preprints: An early career researcher perspective. Sarabipour et al. PLoS biology (2019)[.

Authors’ response:

Updated.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Citations

    1. Chiarelli A, Johnson R, Pinfield S, et al. : Interviewees and mapping of interview questions to areas of Innovation Diffusion Theory.2019a. 10.5281/zenodo.3538919 [DOI]

    Data Availability Statement

    Underlying data

    The authors confirm that, for approved reasons, some access restrictions apply to the data underlying the findings. Data underlying this study cannot be made publicly available in order to safeguard participant anonymity and that of their organisations. Ethical approval for the project was granted on the basis that only extracts of interviews would be shared (with appropriate anonymisation) as part of publications and other research outputs. In order to share data with other researchers, the participants must be contacted and consent to this data being released. In order to request data release, other researchers should contact the corresponding author or Chair of the University of Sheffield Information School Research Ethics Committee ( ischool_ethics@sheffield.ac.uk).

    Extended data

    Zenodo: Interviewees and mapping of interview questions to areas of Innovation Diffusion Theory, http://doi.org/10.5281/zenodo.3538919 ( Chiarelli et al., 2019a).

    Extended data.csv contains our interview questions, split by stakeholder group and mapped to innovation diffusion theory. List of interviewees.csv contains the names, roles and affiliations of interviewees.

    Extended data are available under the terms of the Creative Commons Zero “No rights reserved” data waiver (CC0 1.0 Public domain dedication).


    Articles from F1000Research are provided here courtesy of F1000 Research Ltd

    RESOURCES

    close