Keywords
biomedicine, open science, open science practices, preprinting, preprints, researchers
This article is included in the Research on Research, Policy & Culture gateway.
Preprints are scientific manuscripts that are made available on open-access servers but are not yet peer-reviewed. Although preprints are becoming more prevalent, uptake is not optimal. Understanding researchers’ opinions and attitudes toward preprints is valuable to optimize their use. Understanding knowledge gaps and researchers’ attitudes toward preprinting can assist stakeholders, such as journals, funding agencies, and universities, to use preprints more effectively. Here, we aimed to collect perceptions and behaviours regarding preprints across an international sample of biomedical researchers.
Corresponding authors of articles published in biomedical research journals were identified from a random sample of journals from the MEDLINE database. Their names and email addresses were extracted to invite them to our anonymous, cross-sectional survey, which asked participants questions about their knowledge, attitudes, and opinions regarding preprinting.
The survey was completed by 730 respondents providing a response rate of 3.20% and demonstrated a wide range of attitudes and opinions about preprints with authors from various disciplines and career stages worldwide. Most respondents were familiar with the concept of preprints but most had not previously posted one. The lead author of the project and journal policy had the greatest impact on decisions to post a preprint, whereas employers/research institutes had the least impact. Supporting open science practices was the highest ranked incentive, while increasing authors’ visibility was the highest ranked motivation for publishing preprints.
Although many biomedical researchers recognize the benefits of preprints, there is still hesitation among others to engage in this practice. This may be due to the general lack of peer review of preprints and little enthusiasm from external organizations such as journals, funding agencies, and universities. Future work is needed to determine optimal ways to improve researchers’ attitudes through modifications to current preprint systems and policies.
biomedicine, open science, open science practices, preprinting, preprints, researchers
The title of the manuscript remained unchanged. However, the abstract has been updated to provide a more neutral framing of the study, ensuring that the discussion of preprints and peer review is balanced and supported by the study data, rather than presenting a preconceived notion in favor of preprinting. Regarding the author list, no changes were required. In the methods section, we revised the sampling framework to address concerns about journal selection. We clarified the use of filters from the NLM catalog to ensure that only journal entries were included, and provided a list of the 400 journals used as supplementary data. Additional clarifications were made regarding the expected response rate, citing literature to support our initial prediction and addressing the reviewer’s request for more context on ResearchGate's inclusion as a preprint platform. We also added a reference for the MedRxiv preprint and revised citations to conform with F1000Research’s citation guidelines for web links. In the results section, Table 3 has been updated to include percentages, improving the consistency of data presentation across all tables. Moreover, we re-emphasized the unfamiliarity of journal policies regarding preprints, rather than attributing the lack of preprinting behavior solely to unsupportive journal policies, as initially suggested. Lastly, in the discussion, we addressed concerns about the framing of peer review and preprints. Changes were made to the introduction and background to better contextualize the survey results, highlighting concerns around peer review and preprinting in a balanced manner. The strengths and limitations section was also updated to reflect limitations regarding the dominance of clinical researchers in the sample.
See the authors' detailed response to the review by Artemis Chaleplioglou
See the authors' detailed response to the review by Kimberly Powell
The term “preprints” refers to scientific manuscripts that are made available on an open-access infrastructure, known as the “preprint server,” which has not yet undergone formal peer review.1 Preprints are typically posted by authors at a point in time before a manuscript is peer reviewed by a journal. Peer review is often considered necessary for publishing scholarly work. However, the time lag from journal submission to publication, typically takes around 8 to 10 months in the biomedical sciences due to multiple review rounds, editorial decision-making, high rejection rates, and time needed for revisions.1–5 Preprints do not rely on peer review prior to release, allowing knowledge to be shared quickly with the introduction of novel results and methodology that can save months to years of research time.4 While preprints may face scrutiny for not undergoing peer review, it is important to note that peer review is not always effective.6 Preprints have become valuable for authors, as a larger group of researchers can critique their work (i.e., open peer review) via the preprint servers’ feedback system that allows for public and open feedback directly onto a manuscript that encourages discussion.7 The focused, appropriate, specific, and transparent (FAST) principles are guidelines that reviewers can follow when reviewing preprints, which allow for high-quality, constructive feedback to be provided.7 As a result, researchers can receive helpful suggestions that would not be provided by journals and therefore, make improvements to their final manuscript, so it can possibly be published sooner and with fewer revisions.2 In addition, authors have also begun to cite preprints more frequently in their working manuscripts as institutions and funders have become more open and permitting in their policies regarding preprints.2 Furthermore, funders acknowledge the importance of preprints by encouraging researchers to post and reference preprints in their grant applications.2 Preprint servers still utilize a minimal screening process to evaluate articles for incompleteness, plagiarism, and whether they blatantly disregard or contradict widely accepted medical practices that could potentially result in harm to others if posted.8
However, despite their benefits, many researchers are hesitant about preprinting. Major concerns include the credibility of preprints and premature media coverage.2,5,9 During the COVID-19 pandemic, many media outlets used preprints to rapidly introduce novel research to the public. However, patients and the public may be less familiar with the fact that preprints have not yet undergone peer-review and may not understand how to discern the difference between a preprint and an article published in peer-reviewed literature.2,9 Nonetheless, much COVID-19 research posted on preprint servers has received high levels of coverage, and certain preprints could be used to push certain agendas, such as conspiracy theories and xenophobia,8 thus leading to the spread of misinformation.9,10
The purpose of this study was to explore biomedical researchers’ attitudes toward preprinting through the administration of an international, anonymous, online cross-sectional survey. We aimed to determine the factors that influence biomedical researchers’ opinions about posting and viewing preprints.
Prior to commencing this study, a protocol was added to the Open Science Framework (OSF) prior to participant recruitment.11 The MEDLINE database was used to identify participants using publicly available information. Once identified, the participants were invited to complete a survey online. Survey questions were purpose-built by a core team of authors. They focused on understanding biomedical researchers’ attitudes and opinions concerning preprinting. This study was approved by the Ottawa Health Sciences Research Ethics Board (OHSN-REB Number: 20220584-01H) on November 17, 2022.
The protocol, study materials, and raw and analyzed data have been made available via OSF at the time of publication submission.11 A preprint of our manuscript has been posted on MedRxiv.12
We conducted a large-scale, anonymous, online, cross-sectional survey of authors who had published in biomedical journals contained within the NLM catalog.
The sampling method followed that as described by Ebrahimzadeh et al.13 We downloaded the MEDLINE database that contained approximately 30,000 journals. The “Journals currently indexed in MEDLINE” filter from the NLM Catalog was used to ensure only journals were included in the study. From this list, 400 journals (Extended data: Supplementary File 1) were randomly selected using the RAND() function of Microsoft Excel for Microsoft 365 MSO (Version 2310 Build 16.0.16924.20054) (2021). Authors’ names and emails were extracted using our script from articles published from 2021/07/01 to 2022/08/01. Extended data: Supplementary File 2 outlines the semi-automated process we used, generated by Ebrahimzadeh et al.13,14
Only researchers who were identified by our sampling framework developed by Ebrahimzadeh et al.13 received a closed survey. Using SurveyMonkey, the prospective participants received an email invitation to complete the survey. The email included an authorized recruitment script that described the study and its goals, invited recipients to review an informed consent form, and asked them to complete an anonymous online survey. Upon clicking on the survey link in the invitation email, participants had to read and agree with the informed consent form before being able to see the survey. The initial list of 24 000 names and e-mails of the corresponding authors contained duplicates and no longer functioning e-mails. We were able to determine that functioning emails were no longer available due to SurveyMonkey’s bounced function, which labels certain contact information that has a permanent reason for an email or text to not be delivered. Before recruitment, duplicate authors and no longer functioning emails were deleted from the dataset. Therefore, 22 808 corresponding authors were emailed. Participants were not offered any financial compensation, nor were they required to participate. Additionally, participants could skip any questions that they did not wish to answer.
The complete survey is provided in the Extended data: Supplementary File 3.14 The first draft of the survey was created by JYN, then pilot tested by SEP, DM, KC, and ACA; feedback from the pilot test was discussed by the team and led to a final version of the survey. The survey was then distributed, and responses were collected both using the University of Ottawa’s approved version of SurveyMonkey. The survey contained 29 questions, and SurveyMonkey estimated that it would take approximately 15 minutes to complete. The survey initially asked participants five general demographic questions including gender, research role, area of research, and country of residence. Participants were then asked nine questions about their experiences with preprints, followed by fifteen questions about their preferences and opinions regarding preprinting. The survey used adaptive formatting, whereby a participant’s prior response would influence the next question they see and the answers that might be provided. Most of the questions were multiple-choice, while a handful were open-ended. Two reminder emails were sent to the participants, each spaced one week apart after the original invitation email was sent (i.e., three emails in total). The survey was closed approximately five weeks after the initial invitation email was sent. Both the initial invitation and reminder emails clearly made mention of this to the survey invitees. Responses were collected from January to February 2023.
In accordance with Ebrahimzadeh et al.,13 survey data gathered from the participants were exported to Microsoft Excel. Basic descriptive statistics, such as counts and percentages, were generated based on the analysis of the quantitative data. Based on gender, employment sector, current career stage, and research area, information and replies from participants were compared and studied. Thematic content analysis was used by different members of the research team to individually code replies to specific qualitative items. To reach an agreement on the respective codes, which were categorically classified and specified into distinct tables, researchers carried out multiple rounds of discussions. The Checklist for Reporting Results of Internet E-Surveys (CHERRIES) was used to report this survey.15
A total of 730 responses were received; the raw response rate is 3.04%, the response rate excluding the bounced emails is 3.20%, and the completion rate is 95.48%. Incomplete responses were defined as those with no questions answered on the second page of the survey onwards. Moreover, it should be noted that this response rate is underestimated because we cannot determine how many of the 22 808 authors who were emailed identified as biomedical researchers. It was also impossible to know a number of external factors that would impact the response rate, such as authors who moved institutions, were on vacation, were ill, and passed away, and thus, were unable to respond. As all questions were optional to answer, we therefore provide the total number of respondents for each question. The percentages provided were calculated using each question’s total number of responses. Over three-fifths of respondents (n=455, 62.41%) were identified as senior researchers, which we defined as researchers who started their careers after formal education over ten years ago. Of the 729 responses, participants were identified as follows: academia (n=580, 79.56%), research staff with no formal academic or industry position (n=58, 7.96%), government scientists (n=28, 3.84%), third sector (n=8, 1.10%), and pharmaceutical industry (n=5, 0.69%); none of the participants were part of the scholarly communication industry. We received 50 responses (6.86%) in which participants chose the “other” option, which mainly consisted of participants from the biotechnology industry, retired personnel, clinical researchers, clinicians, medical practitioners, and students. Detailed demographics and other aggregate participant data are presented in Table 1. The responses from participants were also stratified based on gender, employment sector, current career stage, and research area.
Gender (n=728) | |
Male | 401 (55.08%) |
Female | 312 (42.86%) |
Prefer not to indicate | 14 (1.92%) |
Non-binary | 1 (0.14%) |
Country of primary affiliation (n=698) | |
United States of America | 189 (27.08%) |
Canada | 69 (9.89%) |
India | 41 (5.87%) |
United Kingdom of Great Britain and Northern Ireland | 38 (5.44%) |
Australia | 34 (4.87%) |
Spain | 32 (4.58%) |
China | 29 (4.15%) |
Italy | 25 (3.58%) |
Brazil | 18 (2.58%) |
Netherlands | 18 (2.58%) |
Iran (Islamic Republic of) | 16 (2.29%) |
Turkey | 14 (2.01%) |
Germany | 12 (1.72%) |
Switzerland | 12 (1.72%) |
France | 11 (1.58%) |
Denmark | 9 (1.29%) |
Greece | 9 (1.29%) |
Mexico | 8 (1.15%) |
Japan | 7 (1.00%) |
Norway | 7 (1.00%) |
Austria | 6 (0.86%) |
Republic of Korea | 6 (0.86%) |
Belgium | 5 (0.72%) |
Colombia | 5 (0.72%) |
Poland | 5 (0.72%) |
Romania | 5 (0.72%) |
Sweden | 5 (0.72%) |
Indonesia | 4 (0.57%) |
Portugal | 4 (0.57%) |
Russian Federation | 4 (0.57%) |
Thailand | 4 (0.57%) |
Chile | 3 (0.43%) |
Ireland | 3 (0.43%) |
Pakistan | 3 (0.43%) |
United Arab Emirates | 3 (0.43%) |
Albania | 2 (0.29%) |
Bangladesh | 2 (0.29%) |
Czech Republic | 2 (0.29%) |
Jordan | 2 (0.29%) |
Malaysia | 2 (0.29%) |
New Zealand | 2 (0.29%) |
Other* | 17 (2.38%) |
Current career stage (n=729) | |
Senior researcher | 455 (62.41%) |
Mid-career researcher | 139 (19.07%) |
Early career researcher | 89 (12.21%) |
Other (please specify) | 25 (3.43%) |
Graduate student | 21 (2.88%) |
Current employment sector (n=729) | |
Academic | 580 (79.56%) |
Research staff (no formal academic/industry position) | 58 (7.96%) |
Other (please specify) | 50 (6.86%) |
Government scientist | 28 (3.84%) |
Third sector (e.g., NGO, non-profit) | 8 (1.10%) |
Pharmaceutical industry | 5 (0.69%) |
Current research area (n=727) | |
Clinical research | 335 (46.08%) |
Other (please specify) | 125 (17.19%) |
Pre-clinical research - in vivo | 77 (10.59%) |
Health systems research | 74 (10.18%) |
Pre-clinical research - in vitro | 72 (9.90%) |
Methods research | 44 (6.05%) |
Familiarity with preprints
We asked how familiar participants were with the concept of a preprint based on the Framework for Open and Reproducible Research Training (FORRT) definition, “A publicly available version of any type of scientific manuscript/research output preceding formal publication”. Of the 694 responses, participants ranked how familiar they were with the concept of preprints: very familiar (n=167, 24.06%), familiar (n=269, 38.76%), and somewhat familiar (n=92, 13.26%). Only 61 (8.79%) respondents had never heard of this term.
Publishing experience
We then asked about the participants’ personal experiences with posting on preprint servers. Of the 694 responses, approximately one-third (n=237, 34.15%) had authored more than 31 publications in the past five years. Most participants had not posted a manuscript on a preprint server, with 540 (78.15%) participants of 691 responses not posting their most recent work on a preprint server, and 414 (59.74%) respondents of 693 responses having never posted a manuscript on a preprint server. However, when asked if participants would create a preprint in the future, out of 695 responses, 303 (43.60%) participants said they would not, 296 (42.59%) participants said they were unsure if they would, and 96 (13.81%) participants said they would create a preprint in the future.
Of the 335 respondents who had posted preprints before, the most common preprint server used was bioRxiv, with 148 (44.18%) respondents having published previously on this server. In addition, we asked at what point in the publication process were preprints posted, and of the 294 respondents who had previously posted preprints, 110 (37.41%) participants posted their preprint prior to submitting them to a journal, while 133 (45.24%) participants posted a preprint simultaneously while submitting to a journal.
Viewing/downloading preprints
Interestingly, although most participants did not publish their own works on preprint servers, more than two-thirds of the 695 respondents (n=481, 69.21%) had previously viewed or downloaded preprints, although only approximately a quarter (n=176, 25.32%) had cited a preprint. The most common preprint servers used to view/download preprints among 682 responses were bioRxiv (n=243, 35.63%), MedRxiv (n=148, 21.70%), and ResearchGate (n=140, 20.53%). It is important to note that unlike bioRxiv and MedRxiv that are exclusively preprint servers, ResearchGate is not perceived by all to be a preprint server, as it serves other functions, such as asking and answering questions, connecting/networking with other scientists, and finding collaborators. Therefore, as we did not ask how participants interact with ResearchGate, it is possible that participants selected it for reasons other than viewing preprints.
Peer reviewing preprints
We asked whether peer review should become a part of the preprinting process, and 684 responded, with 231 (33.77%) respondents believing that peer review should be part of the preprinting process, while 240 (35.09%) respondents said it should not. In contrast, 213 respondents (31.14%) were unsure (Table 2). We also asked if our participants had experience with peer reviewing a preprint, and from the 688 responses, the vast majority (n=638, 92.73%) stated they had not. Of those who had peer-reviewed a preprint, we wanted to know if the FAST principles were used during the process, and we received 205 responses. Of the respondents, 148 (72.20%) stated that they were not familiar with the FAST principles, 44 (21.46%) said they did not use the FAST principles, and 13 (6.34%) said they did.
Yes | 231 (33.77%) |
No | 240 (35.09%) |
I am not sure | 213 (31.14%) |
We also asked if participants believed that patients or members of the public should be able to peer-review preprints. Of the 445 who expressed an opinion, 320 (71.91%) believed that patients and members of the public should not be able to peer review preprints, while 125 (28.09%) believed that they should have that opportunity.
We then aimed to determine the factors that impact an author’s decision to post a preprint. We asked participants to assess which had the most significant impact on posting a preprint: employer/research institution, funding agency, lead author, co-author consensus, and journal policy. From the 651 respondents, the lead author had the most significant impact on the decision to post a preprint, followed by journal policy, co-author consensus, and funding agency, with the least impactful being the employer/research institution (Figure 1).
We also asked whether employers required or prohibited the posting of a preprint prior to submission to a journal. Most employers did not, with 649 (94.06%) respondents of 690 responses stating that their employer does not require them to post a preprint, and 657 (95.08%) respondents of 691 responses stating that employers do not prohibit them from doing so. In addition, we wanted to see if a funder may impact this decision, and of the 692 responses, 639 (92.34%) participants stated that funders do not require them to post a preprint.
We also wanted to determine how familiar participants were with the preprint policies of journals that they had published with previously. First, we asked whether participants were familiar with either Sherpa Romeo or Transpose Publishing, which are resources that present the preprint policies of most peer-reviewed journals. Of the 686 responses, 623 (90.95%) were not familiar with Sherpa Romeo, and 656 (95.91%) were not familiar with Transpose Publishing. In addition, we asked what the preprint policy was for the journals of respondents’ most recent first author/co-author publications (Figure 2). Of the 684 responses, 282 (41.23%) stated that they were not familiar with the preprint policies, while 183 (26.75%) stated that preprinting was permitted. Only 34 (4.97%) respondents and 61 (8.92%) respondents stated that preprinting was prohibited or that they had no preprinting policies.
Respondents were asked to what extent they agreed with certain incentives or the consequences of preprinting (Figure 3). Of the 666 who answered, over half had positive attitudes toward the following incentives and consequences: 1) preprints support open science practices (n=456, 68.99%); 2) preprints should not be cited because they may contain results that lack credibility because they had not yet undergone peer review (n=433, 65.12%); 3) preprints can be used as a weapon for the dissemination of misinformation (n=417, 62.61%); 4) preprinting behaviors will increase in the biomedical field in the future (n=415, 62.32%); and 5) preprinting allows for more efficient scientific dissemination (n=400, 60.06%). In terms of whether respondents agreed or disagreed with the statement “Preprinting may encourage other researchers to steal project ideas (i.e., scooping), less than half (n=304, 45.66%) respondents agreed, 143 (21.47%) respondents remained neutral, and 219 (32.89%) respondents disagreed. On the other hand, 284 respondents (42.71%) did not agree that preprinting would ultimately lead to higher-quality research being published, while 224 (33.68%) and 157 (23.60%) respondents chose neutral and agreement options, respectively.
We then asked the extent to which respondents agreed with the following motivations for publishing a preprint (Figure 4). Of the 666 respondents who answered, over half had positive attitudes toward the following incentives and consequences: 1) preprinting increases authors’ visibility and allows for networking (n=489, 73.43%); 2) preprinting allows for early/more feedback on their manuscript submission (n=470, 70.58%); 3) preprinting allows authors to share studies that have been rigorously conducted but present negative/perceived low-impact results (i.e., lower likelihood of publication in a peer-reviewed journal) (n=437, 65.70%); 4) preprinting makes my research visible to all if it is published in a paywall journal (n=415, 62.89%); and 5) preprinting increases the likelihood of authors receiving more views and citations on their published article (n=365, 55.14%). In addition, slightly less than half of respondents (n=323, 48.57%) agreed that preprinting allows authors to prevent duplication of efforts, while 193 (29.02%) disagreed, and 149 (22.42%) remained neutral. In terms of if respondents agreed or disagreed with the statement “preprints aid researchers’ careers with respect to hiring, promotion, and tenure”, respondents’ opinions were more evenly split, with 205 (30.91%) respondents agreeing, 218 (32.88%) respondents remaining neutral, and 240 (36.19%) respondents disagreeing.
In addition, we asked participants for their opinions on making preprinting mandatory during the research process. Of the 665 respondents who expressed opinions, over half (n=363, 54.58%) were leaning toward disagreeing with the statement. Only approximately a quarter (n=163, 24.52%) of the respondents had positive opinions regarding this statement.
We then asked an open-ended question to allow respondents to freely express other reasons that motivated or dissuaded them from preprinting (Table 3). The 197 responses received could be primarily split into positive and negative factors. The most prevalent positive factor that motivated preprinting “preprints allow rapid dissemination”. This was followed by “preprints show research efforts,” in which preprinting allowed authors to show their research to other parties, such as other researchers, funding agencies, and universities, without the use of a publication in a journal. Lastly, the final positive factor commonly mentioned was “preprints are open-access items”.
On the other hand, the open-ended questions showed that many negative factors discouraged respondents from preprinting. The most frequent, with approximately a third of responses, was “peer review is bypassed,” as many responses stated that peer review must take place to determine the credibility and quality of research before publishing. In addition, many respondents believed that “preprinting is not needed,” since preprinting does not provide any benefits to the individual’s career and there were no issues with the current publishing system. Another negative factor was “preprints might release misinformation,” followed by “preprints are poor-quality research.” Next, many responses stated that the “journal policy prevented preprinting”, since journal policy rejected or was unclear about preprinting. “Scooping” was also mentioned frequently, along with “preprinting requires additional resources,” where respondents felt that preprinting needed extra time, effort, and funding that is deemed unnecessary for the overall publication of their work. Additionally, many respondents were “unfamiliar with preprints” which discouraged them from preprinting. Lastly, “multiple versions of the same paper” was another negative factor, where having a preprint and a journal published version of the same paper results in the release of unedited, possibly incorrect, work and causes confusion when referencing.
The purpose of this study was to explore biomedical researchers’ attitudes toward preprinting through the administration of an international and anonymous cross-sectional online survey. We found that the most impactful factors that turned authors away from preprinting were the fear of disseminating misinformation, lack of peer review, and unfamiliarity with the preprinting policies of journals. Respondents believed that preprints are beneficial in increasing visibility and recognition, being openly accessible, providing rapid feedback, and publishing negative data.
Future work might benefit by exploring strategies for improving researchers’ attitudes toward preprinting and examine how other stakeholders can adjust their preprint policies in the future. To the best of our knowledge, this cross-sectional survey is the first to specifically investigate the attitudes of biomedical researchers toward preprinting. The insights gained from this study are crucial for understanding how biomedical researchers view preprints, which could influence the development and implementation of preprint policies by researchers, institutions, and publishing houses. In addition, this study may be used as a model for others interested in preprinting practices and polices within their discipline of study.
In terms of the benefits and consequences of preprints, the three statements that respondents most agreed with were: 1) preprints support open science practices, 2) preprints should not be cited because they may contain results that lack credibility because they have not yet undergone peer review, and 3) preprints can be used as a weapon for the dissemination of misinformation. These results align with the preliminary work done by Funk et al., who found that the most concerning issues with preprints from the perspective of respondents, who were mainly researchers in academia, were premature media coverage and public sharing of information prior to peer review, while respondents believed that the greatest benefit of preprints is making their work open access.
On the other hand, the top three most impactful factors among respondents with respect to deciding to publish a preprint included 1) increasing author visibility and networking opportunities, 2) allowing for early and more feedback on manuscripts, and 3) publishing of negative or perceived low-impact results. The findings by Funk et al. somewhat align with our results, as they found that only a little over half of the respondents found early feedback as “very beneficial” and were only seen as less impactful in comparison to open access and speed of dissemination. However, publishing negative or perceived low-impact results was the second least impactful factor, only being seen as “very beneficial” by approximately one-third of the respondents. This variation may be due to greater awareness of the bias of journals toward positive results, and why this is an issue for the scientific community. Nowadays, authors are pressured to only publish positive “groundbreaking” results, and journals often reject papers with negative or non-significant results to increase readers, citations, and submissions.16–19 However, negative and non-significant results are vital to the scientific process as they allow for collective self-correcting progress and ensure that resources are not wasted on the replication of failed research.17,18 A survey by Echevarría et al.19 showed that the majority of authors believe this information should be shared, although only a handful have tried to publish negative results. Preprints may be an appropriate alternative for publishing ‘negative’ data by allowing researchers to quickly disseminate their findings without delays of a lengthy peer review process. Also, unlike traditional journals, preprints eliminate the risk of manuscript rejection and avoid concerns related to journal metrics. Future work should focus on gaining a better understanding of researchers’ attitudes toward publishing negative results and preprint practices to ensure that negative results are appropriately disseminated within the scientific community.
Scooping is often a negative factor that impacts authors’ decision to share their work with open access, at conferences, and on preprint servers.20 Scooping is when someone unethically claims a research idea or results as their own via earlier publishing and often occurs due to public sharing of research prior to publishing in a journal.20,21 As a result, scooping can prevent researchers from publishing their work in journals and turns their research into a waste of effort.20 In our survey, slightly less than half of our respondents believed that preprints encouraged scooping, and this was also mentioned as a concern against preprinting in the open-ended questions. Thus, our findings agree that perceived scooping prevents researchers from publishing in preprint servers. However, there is no evidence of scooping via preprints occurring or that preprints encourage scooping, as observed by Tennant et al.21 Rather, preprint servers actually protect against scooping, as preprint servers provide time-stamps and establishes priority of discovery, which does not occur when presenting research in public.21 Efforts should therefore be made to show researchers that scooping and its consequences are unlikely to occur, and instead, researchers should focus on the benefits of preprinting.
One hesitation toward the use of preprints revolves around the fact that preprints have not undergone peer review; therefore, many researchers believe the manuscript content has not been validated before being released to the public. This is seen in our findings, whereby almost two-thirds of our respondents believe that preprints should not be cited because the results may lack credibility since there was no peer review. Moreover, over two-thirds of our respondents had viewed or downloaded a preprint, but only a quarter had ever cited it, implying distrust in preprint content. Additionally, peer review being bypassed was the most frequent factor that negatively impacted respondents’ motivations to preprint, as seen in the open-ended question, alongside the fear that preprints will disseminate misinformation and poor-quality research. Thus, we consistently observed that peer review is regarded as highly important and necessary for publishing scholarly work by respondents. Respondents’ hesitation toward preprints due to lack of peer review and fear of misinformation indicated a lack of knowledge about the evidence regarding the quality of the published literature and the effectiveness of peer review. Research conducted by Zeraatkar et al.22 has shown that there was no evidence that preprints provided results that were inconsistent with peer-reviewed publications, specifically seen with COVID-19 treatments. Although Bai et al.23 found differences in results between preprints and publications, the main conclusions remained consistent.
MEDLINE is also currently indexing preprints as part of the NIH Preprint Pilot project, supporting research that has been posted to eligible preprint servers for easy discoverability and preservation. In addition, it is important to note that peer review is not always effective because of inconsistency and bias, as many peer reviewers have not received formal training on how to effectively peer review, and many training opportunities and courses are not openly accessible online, inhibiting researchers from gaining this education.22,24–27 Thus, many believe that the peer review process ensures quality research is published may negatively impact researchers’ opinions toward preprints.
Interestingly, despite many respondents distrusting preprints due to a lack of peer review, many were divided on whether peer review should be incorporated into the preprint process. This may be because preprints allow for the rapid dissemination of research as there is a delay from journal submission to publication when undergoing peer review.1,5 If peer reviews were to be incorporated into preprinting in the same way as journals, then the work of peer reviewers would increase, and the delay would become much longer. It has thus been suggested by Soderberg et al.,5 that preprint servers should be modified to improve the credibility of preprints, as the current preprint feedback system allows for public feedback and discussion, which may aid in credibility or results. However, authors often fear unfair criticism from competitors, harm to their reputation, or softened criticism due to the public nature of feedback.8 The FAST principle may be able to alleviate these issues.7 Based on our findings, the majority of respondents had not peer-reviewed a preprint before, and of those that had, most were not familiar with the FAST principles. Therefore, efforts should be made to educate researchers on these principles to promote feedback on preprints. Future work may aim to better understand researchers’ opinions on peer reviewing preprints and potentially provide other novel solutions such that the benefits of peer review can be implemented in the preprinting process while simultaneously allaying the current fears of preprint feedback.
Journal policy is the second most impactful factor in the decision of authors to post a preprint and was frequently mentioned in the open-ended question. This may be a result of unfamiliarity towards a journal’s preprint policy, as seen in the survey. Nowadays, most, though not all, life science journals accept preprinted manuscripts for submission, and some journals either do not have a preprint policy or have contradicting statements.3 Biomedical journals should create clear and concise preprint policies that allow researchers to understand whether they can publish preprints before or during submission to a journal. In addition, our data showed that most researchers were unfamiliar with resources about preprint policies, such as Sherpa Romeo and Transpose Publishing, which gather journal policies on open access and preprinting. In the future, it would be best to resolve the issue of lack of awareness towards journal’s preprint policies by educating researchers on these types of resources to address the concerns that preprints negatively impact the chances of publishing in a peer-reviewed journal.
The respondents did not have strong opinions on whether preprints aid researchers’ careers with respect to hiring, promotion, and tenure. In addition, our findings showed that universities and research institutes had the least impact on respondents’ decision to publish a preprint, and the majority of our respondents’ employers neither prohibited nor permitted the posting of preprints. This may be because few universities policies consider preprints when making decisions regarding hiring, promotion, and tenure. It has been suggested that hiring, promotion, and tenure policies should be modified, as current methods rely heavily on quantitative metrics, which have been recognized as a flawed system.28,29 Preprints may prevent the hiring committee’s bias toward or against a journal due to its impact factor or reputation, as suggested by Polka. The inclusion of preprints in hiring, promotion, and tenure policies may improve biomedical researchers’ attitudes toward preprinting.
By implementing a cross-sectional survey for this project, we took a snapshot of the population of interest without following them over time. This study was able to generalize among biomedical researchers by randomly surveying a large sample of them with varying opinions regarding preprint postings. A limitation of our study design was that it was written in English, so researchers without working knowledge of English could not take the survey. In addition, as the present study was based on an English-speaking international sample, it does not consider the impact of national policies on attitudes and opinions toward preprinting. Another limitation of our sampling strategy is that the list of email addresses used potentially included inactive or invalid addresses, which may have been a result of changing professions, retiring, or passing away; furthermore, we did not consider autoreplies. Therefore, the response rate is underestimated. Our questionnaire was built on self-declared attitudes and preprint practices, which may not accurately capture independently confirmed practices. Inherent to the cross-sectional survey design, the final limitation also includes a low response rate, recall bias (in which participants do not correctly remember past events), and non-response bias (in which participants did not want to or could not complete a survey question). In addition, responses mostly reflected the opinions of senior academics, and there were less responses from early career researchers. Most responses were also from clinical and pre-clinical researchers due to our sampling of the MEDLINE database. Therefore, it is possible that researchers in fields not typically indexed in MEDLINE may have different views and experiences.
The purpose of this study was to explore biomedical researchers’ attitudes toward preprinting through the administration of an international, anonymous, online cross-sectional survey. We observed that biomedical researchers were familiar with the concept of preprints but lacked experience with preprints. The various attitudes and opinions of biomedical researchers provide insights into the important factors that influence the use of preprinting in the scientific community. With this study, we found that respondents believed that preprints have benefits such as increasing visibility and recognition, being openly accessible, providing rapid feedback, and publishing negative data. However, fear of disseminating misinformation, lack of peer review, and unfamiliarity with journal’s preprint policies resulted in authors avoiding preprints. To address preprint hesitation in biomedical researchers, we found that a modified preprint feedback system, clear journal policies regarding preprinting, and recognition of preprints as research contributions could aid in improving researcher attitudes towards preprinting. This study’s findings can inform the provision of suggestions to relevant stakeholders for the implemention and improvement of preprinting.
Written informed consent for publication of the participants’ details was obtained from the participants.
Open Science Framework: Underlying data for ‘An international, cross-sectional survey of preprint attitudes among biomedical researchers’, https://doi.org/10.17605/OSF.IO/QA9GN. 14
This project contains the following underlying data:
Open Science Framework: Extended data for ‘An international, cross-sectional survey of preprint attitudes among biomedical researchers’, https://doi.org/10.17605/OSF.IO/QA9GN. 14
This project contains the following extended data:
Open Science Framework: CHERRIES checklist for ‘An international, cross-sectional survey of preprint attitudes among biomedical researchers’, https://doi.org/10.17605/OSF.IO/QA9GN. 14
Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0).
Views | Downloads | |
---|---|---|
F1000Research | - | - |
PubMed Central
Data from PMC are received and updated monthly.
|
- | - |
Is the work clearly and accurately presented and does it cite the current literature?
Yes
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
Yes
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Yes
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Information Science, Open Research, Research Communication
Is the work clearly and accurately presented and does it cite the current literature?
Yes
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
Not applicable
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Yes
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Health services research, epidemiology, and public health.
Is the work clearly and accurately presented and does it cite the current literature?
Partly
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Partly
If applicable, is the statistical analysis and its interpretation appropriate?
Not applicable
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Partly
References
1. Blatch-Jones AJ, Recio Saucedo A, Giddins B: The use and acceptability of preprints in health and social care settings: A scoping review.PLoS One. 2023; 18 (9): e0291627 PubMed Abstract | Publisher Full TextCompeting Interests: No competing interests were disclosed.
Reviewer Expertise: Research on Research, metascience, scoping reviews, peer review and decision-making, research management processes and practices, open research, qualitative research (observations, netnography). post-award management
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Library and information science (LIS), Bibliometrics, Health information
Is the work clearly and accurately presented and does it cite the current literature?
Partly
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Partly
If applicable, is the statistical analysis and its interpretation appropriate?
Yes
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Partly
References
1. Malicki M, Jeroncic A, Ter Riet G, Bouter LM, et al.: Preprint Servers' Policies, Submission Requirements, and Transparency in Reporting and Research Integrity Recommendations.JAMA. 2020; 324 (18): 1901-1903 PubMed Abstract | Publisher Full TextCompeting Interests: No competing interests were disclosed.
Reviewer Expertise: Meta-research.
Is the work clearly and accurately presented and does it cite the current literature?
Partly
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
Partly
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Partly
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Librarianship; Scholarly Communications; Bibliometrics; Research Impact; Open Access Policies; Predatory Publishing
Is the work clearly and accurately presented and does it cite the current literature?
Yes
Is the study design appropriate and is the work technically sound?
Partly
Are sufficient details of methods and analysis provided to allow replication by others?
Partly
If applicable, is the statistical analysis and its interpretation appropriate?
Yes
Are all the source data underlying the results available to ensure full reproducibility?
Partly
Are the conclusions drawn adequately supported by the results?
Yes
References
1. Avissar-Whiting M, Belliard F, Bertozzi SM, Brand A, et al.: Recommendations for accelerating open preprint peer review to improve the culture of science.PLoS Biol. 2024; 22 (2): e3002502 PubMed Abstract | Publisher Full TextCompeting Interests: No competing interests were disclosed.
Reviewer Expertise: Library and information science (LIS), Bibliometrics, Health information
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | ||||||
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | 6 | |
Version 2 (revision) 04 Nov 24 |
read | read | read | read | read | |
Version 1 03 Jan 24 |
read | read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Already registered? Sign in
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
Comments on this article Comments (0)