ALL Metrics
-
Views
-
Downloads
Get PDF
Get XML
Cite
Export
Track
Research Article
Revised

The transparency of quantitative empirical legal research published in highly ranked law journals (2018–2020): an observational study

[version 2; peer review: 3 approved]
PUBLISHED 07 Mar 2024
Author details Author details
OPEN PEER REVIEW
REVIEWER STATUS

This article is included in the Research on Research, Policy & Culture gateway.

This article is included in the Meta-research and Peer Review collection.

Abstract

Background

Scientists are increasingly concerned with making their work easy to verify and build upon. Associated practices include sharing data, materials, and analytic scripts, and preregistering protocols. This shift towards increased transparency and rigor has been referred to as a “credibility revolution.” The credibility of empirical legal research has been questioned in the past due to its distinctive peer review system and because the legal background of its researchers means that many often are not trained in study design or statistics. Still, there has been no systematic study of transparency and credibility-related characteristics of published empirical legal research.

Methods

To fill this gap and provide an estimate of current practices that can be tracked as the field evolves, we assessed 300 empirical articles from highly ranked law journals including both faculty-edited journals and student-edited journals.

Results

We found high levels of article accessibility (86%, 95% CI = [82%, 90%]), especially among student-edited journals (100%). Few articles stated that a study’s data are available (19%, 95% CI = [15%, 23%]). Statements of preregistration (3%, 95% CI = [1%, 5%]) and availability of analytic scripts (6%, 95% CI = [4%, 9%]) were very uncommon. (i.e., they collected new data using the study’s reported methods, but found results inconsistent or not as strong as the original).

Conclusion

We suggest that empirical legal researchers and the journals that publish their work cultivate norms and practices to encourage research credibility. Our estimates may be revisited to track the field’s progress in the coming years.

Keywords

metaresearch, open science, transparency, credibility, empirical legal research

Revised Amendments from Version 1

In the abstract, we have added the figures suggested by Reviewer 3.

In the introduction, we have reorganised and also distinguished more between error detection type benefits of reproducibility and other possible benefits (Reviewer 1). We have also clarified our discussion of preregistration and questionable research practices (Reviewer 3).

We have enhanced the readability of Figure 1 in line with Reviewer 2’s suggestion.
For the analysis, we have used the code provided by Reviewer 3 both generally and to update Figures 2 and 3 (this also responds to Reviewer 2). These help to improve the readability of those figures. We have also clarified differences between the studies in Table 4 (Reviewer 1).

See the authors' detailed response to the review by Antica Culina
See the authors' detailed response to the review by Stefanie Mueller
See the authors' detailed response to the review by Rob Heirene

Introduction

Increasing the transparency of research is a key component of the ongoing credibility revolution1 occurring in many fields.2 This movement seeks to improve research credibility by ensuring that claims can be tested and critiqued by other researchers. Further benefits of the credibility revolution are efficiency, in that transparent research is reusable by other researchers to explore new questions,3 and that transparent research enhances public trust in science, comporting with lay expectations about how science ought to be conducted.4 Despite its work being cited by courts and policymakers,5 the field of empirical legal research has so far largely refrained from engaging in significant reforms. In this article, we measure the transparency and other related characteristics of 300 empirical legal studies published between 2018 and 2020 in law journals rated highly by traditional metrics. For the purposes of this article, we define empirical research as research that performs analysis on quantitative data.6

The credibility revolution and the role of transparency

The “credibility revolution”7 responded, in part, to a “crisis”8 reported in many fields, in which researchers were unable to replicate the findings of published studies (i.e., they collected new data using the study’s reported methods, but found results inconsistent or not as strong as the original).9 Failures to replicate and other controversies were well-publicized and documented in psychology.10 However, other fields that run adjacent to legal research have not been immune, such as economics11 and criminology.12 Recently, for instance, economists have described and documented reproducibility failures in studies employing secondary data.13

The credibility revolution involves a host of changes to the research process, such as improved transparency, higher standards of evidence, and more replication research.14

Transparency-focused reforms can make research more efficient because other researchers can leverage open data and materials to test new questions, and to synthesize existing data in meta-analyses.15 Conversely, research efforts can be wasted in the absence of open data in the sense that those data cannot be obtained by subsequent researchers seeking to reuse them. This is because researchers change email addresses and institutions or leave academic research behind altogether, making them unavailable to share data upon request.16 Moreover, many researchers who are reachable, decline to share data and materials when they are contacted, or promise to deliver the data but never follow through.17

Transparency and fuller reporting in the form of data sharing, as well as providing more details of methods and statistical analyses performed, allows other researchers to better scrutinize findings and detect errors in research.18 For instance, researchers recently discovered a case of data fraud in a study purporting to find that signing one’s name before versus after providing information in a document reduces dishonesty.19 This study has been cited often for its legal and policy consequences,20 including by the UK Behavioural Insights Team (i.e., Nudge Unit).21 Beyond availability of the raw data, which helped other researchers to uncover the fraud, replication also played a role. Failures to replicate other studies in the paper led to increased scrutiny of the entire set of results, which eventually led researchers to take a closer look at the data. One of the authors of the problematic paper, who had worked on the non-fraudulent studies reported within the same article, wrote in response to the discovery of the fraud:22

Though very painful, this experience has reinforced my strong commitment to the Open Science movement. As it clearly shows, posting data publicly, pre-registering studies, and conducting replications of prior research is key to scientific progress.

Note that this is a quote from Francesca Gino. When we wrote the first version of this article, Gino had not yet been accused of fraud in relation to other studies.23 That second potential fraud attributed to Gino was also discovered by way of the underlying data being available.

In addition to data and analysis scripts (i.e., code that researchers feed into statistical software packages such as R and STATA to produce reported results), transparency is advanced through preregistration (or prospective trial registration and a pre-analysis plan, as it is called in medical research and economics respectively), which is a time-stamped statement of the research protocols and hypotheses that is posted prior to data collection.24 Preregistration is designed to address publication bias (i.e., the tendency for journal editors to prefer studies that produce statistically significant results) and questionable research practices (i.e., practices that increase the likelihood of publication but decrease the likelihood of successful replication—e.g., producing results using many different empirical models and reporting only statistically significant results).

Similarly, registered reports aim to promote transparency and decrease incentives to engage in questionable research practices.25 Registered reports are studies that begin with peer review of the research plan prior to data collection and are accepted or rejected based solely on the plan and whether the researcher, after collecting data, follows the plan, Early research suggests results from studies published using a registered report protocol contain a more realistic proportion of null results.26

Measuring transparency and credibility-related features of published research

Several metascientific studies, across a variety of fields, have conducted “state-of-the-science” audits, in which recent published studies are randomly sampled and coded for various transparency and credibility-related features.27 These metascientific studies have generally found very low levels of transparency. One study examined psychology articles published from 2014-2017.28 Only about 2% of the studies sampled had available data, approximately 17% had available materials, and 3% were preregistered.29 Note, however, that studies published during this timeframe were conducted in the early days of the reported crisis in psychology.30

While these findings are worrisome, recent reforms in other fields may have led to an increase in transparency related practices in recent years. For instance, journals that implemented open data policies (e.g., requiring open data under some circumstances) show substantial increases in the proportion of studies with open data, albeit with imperfect compliance.31

Moreover, a survey across many fields directly asking researchers about when they first engaged in a transparency-related practice (open data, open materials, open code, and preregistration) found that uptake has increased in recent years, suggesting that recent reforms and initiatives are moving the needle.32

Empirical legal research

Numerous researchers have questioned the credibility of empirical legal research. In a relatively early critique, Epstein and King reviewed all law journal articles published over a ten-year period that contain the word “empirical” in the title.33 They found numerous errors, generally centering around poor transparency and reproducibility. For instance, many authors had not fully described how they gathered data and then reasoned from that data to their conclusion. Similar critiques have been levied since then, such as reports that empirical legal studies misinterpret statistical results (e.g., p-values), misapply statistical methods, and fail to verify that the assumptions underlying their methods were met.34 Furthermore, author eminence likely plays a biasing role in empirical legal research because student editors may be especially vulnerable to accepting articles based on the status of the author. Even outside of the student context, author status has been shown to affect peer review decisions.35 Most recently, Huber and colleagues found that an article submitted with a Nobel Laureate as corresponding author received over 40% fewer reject recommendations as compared to the same manuscript with a PhD student as corresponding author.36

Matthews and Rantanen conducted the most recent metaresearch on empirical legal research, measuring data availability.37 They sampled from the top 20 journals in the Washington & Lee rankings from 2010-2019, as well the Northwestern Law Review and the Journal of Empirical Legal Studies. They added the latter two because they provided a contrast with the other journals in the sample in terms of peer review. The Northwestern Law Review is one of the rare student-edited journals to routinely seek peer reviews for empirical work, and the Journal of Empirical Legal Studies is fully faculty-edited and peer reviewed. Matthews and Rantanen found low levels of data availability across the 614 articles in their sample, with only 12% making data available without contacting the author. Moreover – and despite its specialization on empirical works and policy encouraging authors to make their data available – the Journal of Empirical Legal Studies underperformed the other journals with only 6% data availability. These results converge with a 2021 study finding that highly ranked law journals implemented almost no transparency guidelines or requirements.

Limited data availability is especially troubling given several other aspects of empirical legal research that sets it apart from cognate fields. For instance, as individuals formally trained in the law rather than in empirical science, many authors of empirical legal work have less methodological expertise than researchers in other sciences. This lack of training may contribute to errors and unfamiliarity with methodological safeguards. The field’s lack of expertise also limits the usefulness of peer review (for journals that do use it).

These factors suggest that transparency is especially important for empirical legal research. For instance, accessible data and analytic scripts and preregistration can assist with error and bias detection. And, other aspects of transparency, such as articles that are openly available and declare funding sources and conflict of interests, help others assign credibility to reported results. Still, outside of the low data availability at elite journals, there is little current knowledge about transparency of empirical legal research. The last large study that assessed a broad array of transparency indicia was conducted 20 years ago. It included only articles with “empirical” in the title38 and the results were not quantified in a way that makes them easy to update and revisit. This study seeks to fill these gaps.

Methods

Overview and design

To estimate the transparency of credibility-related features of recent empirical legal research, we examined a sample of 300 law journal articles published between 2018 and 2020. We chose this sample size because it is consistent with many previous transparency studies.39 Based on those authors’ reports40 of how long it took them to extract the relevant features of each article, we judged that coding 300 articles was a practical target given our available resources. To provide a comparison between the student-edited journals (that tend to not use peer review, but rather the judgment of student editors to make acceptance decisions) and faculty-edited journals (that tend to rely on peer review), we chose 150 articles from each. We classified articles as empirical if they included original analyses using descriptive or inferential statistics of original or pre-existing quantitative data (e.g., survey studies, content analyses of judicial decisions, meta-analyses).

As described below, we coded features generally related to transparency, such as accessibility, statements about the availability of data, analytic scripts, and other research materials, whether the study was preregistered, and declarations of conflicts of interest and funding sources. We also coded general methodological aspects of those studies, such as whether they were experiments and the types of statistics performed. These provide some background understanding of our sample and may bear on the importance of transparency (e.g., providing analytic code is most relevant to studies using inferential statistics). This is the first study of its kind in empirical legal research, and we are not testing hypotheses; thus, the results should be considered descriptive and exploratory. This study is preregistered and provides open data, code, and materials.

We deviated from previous studies measuring transparency in two main ways. First, previous studies using this type of protocol focused on fields whose journals contain a high proportion of empirical research (e.g., psychology, organizational behavior research, otolaryngology, addiction medicine),41 so they randomly sampled studies without screening out studies that did not use empirical methods. This approach would have been inappropriate for the current study because it would have led us to include a large number of non-empirical studies (~90% of published work, according to a prior estimate).42 As a result, we developed an approach for early screening of non-empirical research (see literature search string below). We also deviated from some previous studies by sampling only from highly ranked journals. This may have biased our results towards finding higher research transparency than the field generally has, because higher rank typically translates to greater selectivity, and thus should in principle enable higher standards. Note also that given the perceived importance of the journals in our sample, low levels of transparency would be especially concerning.

Identifying empirical articles: Search string used to generate sample

To develop a search string to more efficiently identify and sample articles that met our specifications, we conducted a preliminary examination of the literature. We coded 2019-2020 articles from 10 law journals that Washington and Lee ranks in the top 25 (1,024 total articles).43 Through reading those articles, we identified 92 (or 9% of the sample) meeting our definition of empirical within this dataset.44

Using the knowledge from that preliminary examination, we first considered two different ways of more quickly identifying empirical articles without reviewing the full text. First, we considered selecting only articles with the word “empirical” in the title as Epstein and King had done in their landmark study. However, only 10% of the empirical articles in the preliminary examination sample had the word “empirical” in their title. This strategy, therefore, would miss a great deal of empirical work, raising concerns about the representativeness of the sample and making it more difficult to find our target of 300 recent empirical studies. We also considered selecting only articles with “empirical” in their abstract; however, that strategy would have missed approximately 50% of the articles identified by the more intensive method used in our preliminary examination.

Ultimately, we decided to use the words in the abstracts of the 92 empirical articles we identified in our preliminary examination, and to write a search string based on those words. That search string is:

ABS (“content analysis” OR data* OR behavioral OR behavioural OR empirical OR experiment OR meta-ana* OR multidimensional OR multivariate OR quantitative OR statistical OR study OR studies OR survey OR systematic)

One limitation of this strategy is that, in our preliminary examination, about 8% of the empirical articles we identified did not have an abstract. As a result, any search strategy that uses abstract searches is bound to miss a small proportion of empirical articles, such as commentaries with a trivial empirical component. This may bias our findings towards including more instances of systematic data analysis that would be adverted to in an abstract. Despite this limitation, the search method is efficient (i.e., full text searches would have yielded too many false positives for our team to review) and reproducible (i.e., the full search string and results are provided, as are all exclusions and reasons for exclusion).

Sample

Figure 1 details our sampling process and exclusions. We used the search string described above to search Scopus for articles published between 1st January 2018, and the date of our search, 29th January 2021. We populated our overall sample of 300 articles with 150 articles from the top 25 student-edited journals from the Washington and Lee rankings (W&L) (based on its “combined score” in 2019) and 150 articles from the top 25 faculty-edited journals (by 2019 impact factor) in the Web of Science’s “law” database.45 That is, we applied our search string to both of those journal lists. The Washington and Lee search returned 596 articles and the Web of Science search returned 859 articles (see Extended data). We decided to sample from high impact journals because we judged that these articles would be most influential among both researchers and policymakers, and thus transparency is especially important.

a77d7ec5-40c3-48c1-ac50-073d46100aac_figure1.gif

Figure 1. The screening procedure for building the student-edited (W&L) and faculty-edited samples (WoS).

Articles were first identified through the Scopus search string described in the methods. They were then screened for eligibility in random order until the samples were complete. The excluded articles and the reasons for their exclusion are available in the Extended data, “W&L screened out” and "Web of Science screened out”.

Because searches returned several of what we classified as non-empirical articles (e.g., the abstract contained the word “data” to describe data regulation laws), one author (JC) randomly sorted both lists and then screened out articles that did not meet our inclusion criterion (i.e., the study includes an analysis of quantitative data) until we reached the pre-specified sample of 150 articles for each group (Figure 1). Of the 596 articles in the W&L sample, we needed to review 510 to obtain our sample of 150 (i.e., 31% of those reviewed were selected, the rest were excluded). For the Web of Science sample, we needed to review 383 to find 150 empirical articles (i.e., 40.1% of those reviewed were selected, the rest were excluded).

The relatively high rate of exclusions suggests that our search string was overly inclusive, adding more work for us but reducing the chance that we missed a large proportion of empirical articles. The articles screened out and the reasons for their exclusion are described in our Extended data (“W&L screened out” and “Web of Science screened out”). After we initiated coding of these articles with the protocol below, we found that 8 were incorrectly categorized as empirical, so we selected the next 8 from the list as replacements. These are the numbers that are reflected in Figure 1 and above.

Coding procedure

Articles were coded using the structured form developed by Hardwicke and colleagues.46 Following the Hardwicke et al., protocol (as well as other transparency coding projects for systematic reviews, see O’Dea et al.),47 each article was coded by two of the authors, with disagreements resolved through discussion between those coders and a third author if the coders could not agree (see Extended data). The coders were all trained on five articles and did not begin coding the target sample of articles until they reached consensus on the five training articles. As we discuss below, two items were difficult to code, and so we discontinued coding them and do not present the result for them. For multiple-study articles (we defined studies as distinct data collection activities), we coded only the first-reported study. Coding one article in the student-edited sample took about 30-45 minutes. Coding an article in the faculty-edited sample took about 10-20 minutes. This reflects the longer length of the articles in the student-edited sample and that their methods and data were frequently difficult to locate due to the lack of a standard article format. We coded articles from February to September 2021.

The features of the articles that we coded are detailed in the coding sheet and in Table 1 (and further detailed in our preregistration). Some of these features are relevant background information on the studies, such as the statistics used by the researchers, the nature of the data, and data sources. Others are relevant to the transparency and credibility of the research, such as whether authors stated that data and analysis scripts were available, whether the study was preregistered, and whether it was a replication (replications have helped uncover spurious results in prior studies).

Table 1. The primary measured variables in our analysis.

The full set of variables can be found in the full structured coding form.

VariableFurther details
Article accessibility

  • Was the article available through the journal’s website (without university library access, i.e., gold open access)?

  • Was the article available through another service (e.g., ResearchGate, SSRN)?

Conflict of interestDoes the article include a statement indicating whether there were any conflicts of interest?
FundingDoes the article include a statement indicating whether there were funding sources?
Experimental designIs it an experiment? For our purposes, experiments are studies in which some variable is manipulated by the researcher (e.g., some participants are randomly assigned to a condition).
SynthesisIs it a synthesis (e.g., meta-analysis, systematic review)? For our purposes, a synthesis is a quantitative analysis of other studies/articles.
ReplicationDoes the article claim to report a replication study?
Human subjectsWere there human subjects? For our purposes, this means measuring and/or aggregating responses from individuals or groups. This does not include judicial decisions written by judges and analogous data.
Original or secondary dataFor our purposes, original data are data the authors collected or generated that did not exist before. Secondary data are data that already existed (e.g., analyses of judicial decisions or contracts).
Data availability

  • Does the article state whether or not data are available?

  • How does the statement indicate the data are available?

  • Can you access, download, and open the data files (without contacting the author)?

Analysis script availability

  • Does the article state whether or not analysis scripts are available?

  • How does the statement indicate the analysis scripts are available?

  • Can you access, download, and open the analysis files (without contacting the author)?

Materials availability

  • Does the article state whether or not materials are available?

  • How does the statement indicate the materials are available?

  • Can you access, download, and open the materials files (without contacting the author)?

Preregistration

  • Does the article state whether or not the study (or some aspect of the study) was preregistered?

  • Where does the article indicate the preregistration is located?

  • Can you access and open the preregistration?

With respect to data availability, Hardwicke et al. attempted to code whether authors provided a clear reference to where the data could be found (“source of data provided but no explicit availability statement”).48 Due to difficulty coding this item, they did not report this and instead collapsed these types of data references into “no – there was no data availability statement”. Because we expected the current study to include several cases of authors analyzing pre-existing data and datasets, we initially attempted to preserve this as a distinct item in our coding form. However, our coders also encountered difficulty with it (e.g., sometimes articles would provide a vague reference to another article, and, when we accessed that article, it referenced yet other articles). So, our results also collapse these types of data references into the “no data availability statements” category (as we note below, our data availability results are closely in line with Matthews and Rantanen, lending confidence in our data availability conclusions). We did, however, include a separate item for secondary data studies (Table 1) in which we coded whether authors provided an index of the secondary data items (e.g., references to the judicial decisions included).49

We report 95% confidence intervals calculated using the Sison-Glaz method for multinomial proportions.50

Deviations from preregistration

Our study deviated from our preregistration in two ways. First, we originally planned to code sample size but did not complete this coding because studies did not provide a single sample size. Second, as noted above, we originally planned to code whether the authors provided the source of the data, but we did not complete this because it was impractical for reasons noted in the previous paragraph.

Results

Overall, we found a low level of transparency on the characteristics we measured. Only 19% of articles stated that their data are available, and we were able to access that data in only about half of those cases.51 Preregistration and availability of analytic scripts were also very uncommon, and, in fact, almost nonexistent in the empirical legal research examined here. However, we found several positive aspects of the literature to build on. For instance, about 50% of studies employing original data stated that at least some materials were available. In addition, article accessibility was high among the empirical legal research examined here, especially among articles in student-edited journals (100% of those articles were available without library access). These findings are detailed below.

Sample characteristics

General characteristics of our sample are reported in Table 2, specifically the proportion of articles that: analyzed original or secondary data; used human participants; reported an experiment; were a synthesis (which we operationalized as studies that self-identified as a systematic review or meta-analysis); and reported descriptive or descriptive and inferential statistics. Secondary data analysis was more common (65% of studies, 95% CI = [59%, 70%]) than analysis of original data. Secondary data were also more frequently employed in the student-edited journals (79%, 95% CI = [73%, 85%]) than in the faculty-edited journals (51%, 95% CI = [43%, 59%]). Furthermore, 40% (95% CI = [35%, 46%]) of studies relied on human participants. This figure was 21% (95% CI = [15%, 27%]) among the student-edited journals and 60% (95% CI = [53%, 69%]) among the faculty-edited journals.

Table 2. Overview of the samples of empirical legal studies.

The variables are: original or secondary data, whether there were human subjects, whether the study was an experiment, whether it was a synthesis (i.e., systematic review or meta-analysis), and whether it used descriptive statistics or descriptive statistics along with inferential statistics. Studies per year can be found in the markdown file in the online data supplement.

VariableResponseN% [95% CI]
All
Original or secondary dataOriginal10635% [30%, 41%]
Secondary19465% [59%, 70%]
Human subjectsNo17960% [54%, 65%]
Yes12140% [35%, 46%]
Experimental design?No24782% [78%, 87%]
Yes5318% [14%, 22%]
SynthesisNo29498% [97%, 99%]
Yes62% [1%, 3%]
Statistics usedDescriptive9732% [27%, 38%]
Descriptive & Inferential20368% [62%, 73%]
Studies in student-edited journals
Original or secondary dataOriginal3221% [15%, 28%]
Secondary11879% [73%, 85%]
Human subjectsNo11979% [73%, 86%]
Yes3121% [15%, 27%]
Experimental design?No12986% [81%, 92%]
Yes2114% [9%, 20%]
SynthesisNo150100% [100%, 100%]
Yes00% [0%, 1%]
Statistics usedDescriptive6443% [35%, 51%]
Descriptive & Inferential8657% [49%, 65%]
Studies in faculty-edited journals
Original or secondary dataOriginal7449% [41%, 58%]
Secondary7651% [43%, 59%]
Human subjectsNo6040% [33%, 49%]
Yes9060% [53%, 69%]
Experimental design?No11879% [73%, 85%]
Yes3221% [15%, 28%]
SynthesisNo14496% [93%, 99%]
Yes64% [1%, 7%]
Statistics usedDescriptive3322% [16%, 29%]
Descriptive & Inferential11778% [72%, 85%]

Turning to methodology, our sample contained fewer experiments (which require random assignment according to our definition) relative to secondary data analyses (18% of studies, 95% CI = [14%, 22%]). Syntheses were very uncommon, with only six in the sample (all six in the faculty-edited sample). Most articles (68% (95% CI = [62%, 73%])) contained descriptive and inferential statistics (the remaining 32% reported only descriptive statistics). 78% (95% CI = [72%, 85%]) of the faculty-edited articles used inferential statistics versus 57% (95% CI = [49%, 65%]) in the student-edited sample.

Among the 194 articles that used secondary data, 53 or 27% (95% CI = [21%, 35%]) of articles analyzed judicial decisions, 11 (6% (95% CI = [0%, 13%])) analyzed company documents, and a further 11 analyzed statutes or legislation (see “table 2secondary” in Extended data). Human participants were recruited from a variety of groups, with 12 of the 121 articles (10% (95% CI = [2%, 19%])) sampling from university students, 35 (29% (95% CI = [21%, 38%])) sampling from the general population, and 74 (61% (95% CI = [53%, 70%])) sampling from special populations. Those special populations52 included difficult-to-reach groups such as judges, young offenders, and government employees (see “table 2 special” in Extended data).

Article accessibility

The articles in our sample were generally easy to access as compared to estimates from previous metascientific studies in criminology and psychology (Table 3, Figure 2).53 86% (95% CI = [82%, 90%]) of articles had publicly available versions – 100% of the student-edited journal articles and 71% (95% CI = [65%, 79%]) of the faculty-edited group. 70% of articles (95% CI = [65%, 76%]) were gold open access, meaning they were accessible on journals’ websites. This was the case for 100% of the articles in student-edited journals, whereas 41% (95% CI = [33%, 49%]) of the faculty-edited articles were gold open access. Empirical legal researchers also regularly use pre- and post-print services to provide open access versions of their work. 42% (95% CI = [36%, 48%]) of articles in the overall sample were downloadable on SSRN and 22% (95% CI = [18%, 27%]) were downloadable on ResearchGate.

Table 3. Transparency and credibility-related features of empirical legal research.

The variables are: article accessibility, the presence and content (if applicable) of statements about funding, conflicts of interest, data availability, materials availability, and analysis script availability. We further coded whether there was a statement that the study was preregistered and whether the authors described the study as a replication. The figures for materials availability include only the articles that collected original data. Note that this figure reflects availability statements. As discussed in text, actual accessibility was considerably lower.

AllStudent-editedFaculty-edited
VariableResponseN% [95% CI]N% [95% CI]N% [95% CI]
Article accessibilityPaywall only4314% [11%, 18%]00% [0%, 1%]4329% [22%, 36%]
Available25786% [82%, 90%]150100% [100%, 100%]10771% [65%, 79%]
Conflicts of interestNo statement26789% [86%, 93%]14999% [99%, 100%]11879% [73%, 85%]
Conflicts21% [0%, 4%]11% [0%, 2%]11% [0%, 7%]
No conflicts3110% [7%, 14%]00% [0%, 1%]3121% [15%, 27%]
FundingNo statement18060% [55%, 66%]11275% [69%, 82%]6845% [37%, 54%]
No funding21% [0%, 6%]00% [0%, 7%]21% [0%, 10%]
Private4114% [8%, 19%]2718% [12%, 25%]149% [1%, 18%]
Public5619% [13%, 24%]64% [0%, 11%]5033% [25%, 42%]
Public & private217% [2%, 13%]53% [0%, 11%]1611% [3%, 19%]
Data availabilityNo statement24281% [76%, 85%]12483% [77%, 89%]11879% [73%, 85%]
Says available5719% [15%, 23%]2517% [11%, 23%]3221% [15%, 28%]
Not available10% [0%, 5%]11% [0%, 7%]00% [0%, 7%]
Analysis script availabilityNo statement28194% [91%, 96%]14294% [91%, 96%]13993% [89%, 97%]
Says available196% [4%, 9%]86% [4%, 9%]117% [4%, 12%]
Materials availabilityNo statement5956% [46%, 65%]1753% [38%, 71%]4257% [46%, 68%]
Says available4744% [35%, 54%]1547% [31%, 65%]3243% [32%, 55%]
PreregistrationNo statement29297% [96%, 99%]14798% [97%, 100%]14597% [95%, 100%]
Says preregistered83% [1%, 5%]32% [1%, 4%]53% [1%, 6%]
ReplicationNo28996% [95%, 98%]14798% [97%, 100%]14295% [92%, 98%]
Yes114% [2%, 6%]32% [1%, 4%]85% [3%, 9%]
a77d7ec5-40c3-48c1-ac50-073d46100aac_figure2.gif

Figure 2. Article availability, funding statements, and conflict of interest statements in empirical legal research.

The left column includes articles from the student-edited sample and the right column is from the faculty-edited sample. Numbers within bars refer to the number of articles that meet the given standard.

Conflicts of interest and funding statements

Turning to conflicts of interest and funding statements, we found that most articles did not provide any such declaration. In fact, only 11% (95% CI = [8%, 15%]) of articles include a conflicts of interest statement. Conflicts of interest statements were more common in the faculty-edited journals with only one article in the student-edited sample containing such a statement. As to statements of funding sources, 40% (95% CI = [35%, 46%]) of articles contained a statement. Again, such statements appear to be rarer in the student-edited sample (see Table 3).

Data availability

The availability of the data, analysis scripts, and materials in our sample was generally low (Table 3, Figure 3). Just 19% (95% CI = [15%, 23%]) of articles provided a statement that data are available. Of articles with data availability statements, the most common means for sharing data were via a third-party repository (39%, 95% CI = [26%, 53%]), by contacting the author (28%, 95% CI = [16%, 42%], and via a personal or institutional website (21%, 95% CI = [9%, 35%]) (see “tabledatahow” in Extended data). We checked whether the data referenced in the statements were readily available (i.e., whether we could access them without further steps, such as contacting the author). Only about half (53%, 95% CI = [40%, 66%]) were readily available, making the effective data availability rate about 10%. This figure closely corresponds to Matthews and Rantanen’s 12% estimate of data availability (also without contacting authors) at predominantly student-edited journals published from 2010 to 2019.54

a77d7ec5-40c3-48c1-ac50-073d46100aac_figure3.gif

Figure 3. Assessment of transparency and credibility-related characteristics of empirical legal research.

The student-edited sample is reported in the left column and the right column is the faculty-edited sample. Numbers within bars refer to the number of articles that meet the given standard. Data availability, analysis script availability, and preregistration bars include the full sample (150 per group), whereas the bars for materials availability include only the articles that collected original data. Note that this figure reflects availability statements, whereas, discussed in text, actual accessibility was considerably lower.

In the social sciences, much of the move towards providing data availability statements has occurred in the context of psychological research, where original data are often collected. As a result, it may be useful to drill down on articles reporting on original data. Limiting our analysis to these articles (N = 106), we found 29% (95% CI = [22%, 39%]) included a data availability statement, whereas only 13% (95% CI = [9%, 18%]) of articles reporting on secondary data did so (N = 194).

For secondary data, as noted above, we coded the steps authors took to provide information about the dataset. In most cases, authors did not provide any details about the dataset (see “table_secondarySteps” in Extended data). In 26 of the 194 (13%, 95% CI = [8%, 20%]) articles reporting on secondary data, the authors provided an index of the secondary data (e.g., a list of judicial decisions relied on). Several others linked to sources, such as external websites, that were no longer accessible.

Analysis-script availability

Very few studies included a statement about the availability of their analysis scripts (6%, 95% CI = [4%, 9%]). Providing analysis code is especially important when reporting inferential statistics (e.g., to determine the exact statistical test and assumptions the authors used), but of the 203 studies that relied on inferential statistics, only 8% (95% CI = [5%, 12%]) made their code available. Even these figures are somewhat inflated, however, because only for approximately half of the articles with script availability statements could we access the scripts without taking further steps (again, due to dead links and statements indicating that the scripts were available on request).

Materials availability

The materials availability results presented in Table 3 and Figure 3 are limited to studies with original data. We presented them this way because sharing of study materials (e.g., survey instruments, vignettes) is arguably less applicable to analysis of existing data. However, some studies analyzing secondary data do involve useful materials that could be shared, such as the coding sheets used by researchers who tally different sorts of judicial decisions. Of studies that reported on original data, about 44% (95% CI = [35%, 54%]) stated that materials were available. Recall that this figure does not mean that all materials were made available, but rather that authors stated that at least some materials were available. Moreover, we were able to access materials for only 39 of the 47 (83%, 95% CI = [74%, 94%]) studies that stated that materials were available, making the effective material available rate about 37% among studies that report on original data.

Preregistration

Almost no studies reported being preregistered (3%, 95% CI = [1%, 5%]). Of the 8 preregistered studies, we could not access the preregistrations of 2. The purported locations of the 8 preregistrations were: the Open Science Framework (5 studies), the AsPredicted.org registry (1 study), the PROSPERO registry (for syntheses; 1 study), and the Evidence in Governance and Politics (EGAP) registry (which is hosted by the Open Science Framework; 1 study).

Replication

Very few studies stated that they were reporting the results of a replication (4%, 95% CI = [2%, 6%]).

Discussion

Our results suggest that there is ample room to improve empirical legal research transparency. Our hope is that our results encourage researchers in the field of quantitative empirical legal research to move forward in making their work verifiable and reusable. Articles in our sample generally had low levels of transparency and credibility-related characteristics that we measured. These results are not much different than many other fields, as shown in Table 4.55 We identified the studies in Table 4 non-systematically, based on studies we were aware from an informal literature search.

On a more positive note, with respect to article accessibility, empirical legal research performs very well, especially for articles published in student-edited journals. Of course, accessibility without fuller transparency risks readers relying on unverifiable results. Ideally, research should be fully transparent and accessible.

Table 4. A comparison of studies measuring transparency-related factors.

Current studyHardwicke et al., 2020Johnson et al., 2020Culina et al., 2020Hardwicke et al., 2018a
Field(s)Empirical LegalSocial SciencesOtolaryngologyEcologyPsychology
Reform(s)?------Journal guidelinesJournal guidelines
Articles analyzedN300250300346174
Publication years2018-20212014-20172014-20182015-20192015-2017
Article availabilityPaywall only14%54.0%77.7%----
Publicly available86%40.4%22.3%----
Data availabilityNo statement81%80.8%96.7%--22%
Says available19%7.0%2.0%79%*78%
Not available0%0.6%1.3%--0%
N300156151346174
Analysis script availabilityNo statement94%98.7%99.4%----
Says available6%1.3%0.7%27%--
N300156151346--
Materials availabilityNo statement56%89.4%94.5%----
Says available44%10.6%4.8%----
N106151145----
PreregistrationNo statement97%100%95.4%----
Says preregistered3%0%4.0%----
N300156151----
ReplicationNo96%98.7%100%----
Yes4%1.3%0%----
N300156151----

* Culina et al. did not study availability statements, but data availability per se – a fuller description of the methodological differences between these studies and an expanded table is available (Extended data, “Table 4 - online supplement”).

Comparing student-edited and faculty-edited journals on other transparency and credibility-related characteristics, we generally did not find large differences. However, student-edited journals did seem to have a smaller proportion of articles with conflicts of interest and funding statements. Deficiencies in reporting funding may be due to law professors relying largely on internal funding that they do not see as important to report. While such funding might raise fewer concerns than that from external sources, it is impossible for the reader to know – without a statement – whether a study received funding and from what source. The best practice, one we saw among some articles in our sample, would be to explicitly declare funding sources and conflicts or the lack thereof, and law journals should require these declarations. Moreover, many legal researchers may have affiliations that should be disclosed, such as governmental appointments, affiliations with think tanks, and company directorships or board memberships.

While we urge caution in comparing our results to those from transparency studies of other fields, such a comparison may be instructive in some ways (see Table 4). In particular, we did not observe large differences (other than in materials availability, see below) between empirical legal research and other fields. However, the two comparison studies in Table 4 (sampling from social science generally and otolaryngology) did not restrict their samples based on journal ranking,56 whereas our study sampled only from what many would describe as the top journals in the field. It arguably would be reasonable to expect that these journals should be leading the field in producing verifiable and reusable work. Moreover, the other studies focus on articles published in the mid-2010s, and so we might expect stronger adoption of transparency and credibility reforms in our sample. In other words, the results of our study likely provide an optimistic comparison with other fields of research.

Regarding the effects of reforms, Table 4 also contains two comparisons with studies that have sampled only from journals that have implemented transparency and openness guidelines. In particular, Culina and colleagues sampled only from ecology journals that had implemented data and analysis script availability policies (both mandatory guidelines and encouragements).57 In addition, Hardwicke et al., examined data availability of studies published by the journal Cognition, which had implemented a mandatory data availability policy.58 As can be seen in Table 4, recent articles in those journals show markedly higher levels of data and script availability than our study found in empirical legal research. We cannot say what caused the relatively high levels of data and script availability in these journals, but these results suggest journal guidelines may play an important role in reform efforts. However, seeing as Matthews and Rantanen found that the Journal of Empirical Legal Studies underperformed student-edited law journals despite having a policy that encourages data sharing, it seems unlikely that mere encouragements are sufficient.

Our results might be limited in other respects. First, empirical legal research is a multi-disciplinary field, which uses a panoply of methods from several research traditions.59 As a result, some forms of transparency may be less applicable for some methods than for others. We attempted to take this into account by reporting results for some of these practices separately for different types of studies (e.g., reporting materials transparency for studies reporting on original data; reporting analysis script transparency for studies reporting inferential statistics). In this respect, our results may overestimate transparency levels by restricting analyses to only one subset of studies, when in fact the practice would be beneficial for a broader range of studies. For example, many studies reporting on secondary data would nevertheless be more reproducible if they shared materials such as coding sheets used by research assistants who coded legislation or judicial decisions.60

Second, we did not contact authors to determine whether statements that data, materials, or analysis scripts were available upon request would be honored or whether authors of studies that do not mention availability would disclose information upon request. As noted above, however, multiple studies have found that most authors do not provide their data when requested, even when their paper includes a statement indicating that data are available upon request.61 Most recently, Gabelica and colleagues found that authors provided just 7% of 1,792 requested datasets despite the authors indicating that the data were available.62 While some authors may have responded to our requests, relying on author responses is problematic in the long run because researchers retire or otherwise leave academia, leading to a “rapid” decrease of research data availability over time.63 In addition, this method of transparency presents a significant obstacle for third parties who wish to access these artifacts for purposes that the authors may view as not in the authors’ interests (e.g., because the requesters suspect an error in the original article). The importance of posting data, as opposed to promising to make it available upon request, has been recognized by government funders, some of whom require authors of funded studies to post data upon publication.64

Third, we did not attempt to take into account data sharing limits such as privacy and proprietary datasets.65 However, we did code whether any statement was made about data availability, which would have included statements about barriers to sharing data, and we did not find any studies that explained their lack of data sharing in such terms, so this may not have been prevalent. Alternatively, authors simply might not have reported their inability to share the data. Moreover, we attempted to code other means of transparency for secondary data analysis (e.g., indexes of cases relied on) and found that few papers took up any such options. Future metaresearch projects may wish to take a more focused approach, targeting specific empirical legal research methods to better understand their norms and limits related to transparent research and reporting.66

Fourth, our coding is only current as of September 2021. If, for example, articles have since been edited to indicate data availability, our results will not reflect that. While that is unlikely, it is perhaps more probable that some articles were temporarily open access because they had just been released, but have now moved behind paywalls. As a result, our results may overestimate open access, especially among the faculty-edited journals published by commercial publishers.

Fifth, using the impact factor metric for Web of Science to identify faculty-edited law journals may have included journals that some in the empirical legal research community would disagree about as important journals in the field. For instance, the impact factor for the Journal of Empirical Legal Studies resulted in it not being included, despite it being the journal produced by one of the main societies in the field. However, including journals based on our subjective judgment would have introduced bias into the findings. And, our results for data availability closely matched that of Matthews and Rantanen, who did study the Journal of Empirical Legal Studies.

Sixth, our sample is potentially biased. If the studies we initially found to develop our search string are different in important ways from the population of studies, the generalizability of our results is limited. That said, our initial sample is sizeable. It includes nearly 100 studies, which reduces the likelihood that we missed sets of relevant studies that are either more or less transparent than the studies in our sample. The bias, of course, depends on the variability of terms in the population of abstracts. In our view, however, the search string terms fairly represent common empirical legal methods and words used to describe them in the literature (e.g., content analysis, behavioral). This gives us confidence that our results describe, at a minimum, a relevant portion of the empirical legal studies literature.

We also highlight that the mere presence of data, analysis scripts, and preregistration does not mean that associated findings will be reproducible. Systematic research has found that data is often not well documented, making it difficult to reproduce findings.67 Future projects should consider focusing on a smaller number of studies for which some data are available to determine if the results are fully reproducible.68 Similarly, other aspects of research quality, such as whether preregistrations were actually followed, are an important avenue for future research.

Looking forward

Where do we go from here? As we reviewed above, transparency has proven vital in uncovering flaws, limitations, and fraud in published work. We call on journals to adopt policies to increase the transparency of published studiessuch as open data and code.69 Such policies can be augmented by “verification checks” whereby the journal verifies all disclosures and uses the disclosed data and code to verify that the article’s results are reproducible. The American Economic Association, for example, performs third-party verifications on all articles published in its journals.70 This may be especially important for journals that are not commonly peer reviewed, such as student-edited journals, because peer review detects some flaws and errors.71 Even then, however, studies have found that peer reviewers detect just a minority of errors deliberately added to the reviewed studies.72 Only with a high level of transparency can we hope that errors in important studies are likely to be caught, as transparency enables robust post-publication peer review.

The fact that at least some datasets employed in empirical legal research studies are proprietary and cannot be made publicly available should not cause the field to shy away from general data availability requirements. For example, in psychology it is common for privacy issues to preclude data sharing. Journal guidelines in this field sometimes balance privacy and other ethical constraints on data sharing with data availability by asking authors to explain any restrictions in the manuscript and requiring data sharing if such an explanation cannot be provided.73 An example of such a statement is: “The conditions of our ethics approval do not permit public archiving of anonymized study data. Readers seeking access to the data should contact the lead author X or the local ethics committee at the Department of Y, University of Z. Access will be granted to named individuals in accordance with ethical procedures governing the reuse of sensitive data. Specifically, requestors must meet the following conditions to obtain the data [insert any conditions, e.g., completion of a formal data sharing agreement, or state explicitly if there are no conditions].”74 This policy is consistent with TOP guidelines for data transparency (Level II), which require data to be posted to a trusted repository and any exceptions to be explained in the article.75 Editors might also consider requiring authors who use proprietary data to include explicit statements related to limitations that arise from the inability to verify claims derived from such data. Specifically, readers should be explicitly warned about relying on unverifiable results.

Ideally, incentive structures for researchers should reward transparency and reproducibility. This includes the research assessment involved in hiring and promotions.76 Research funders should also promote transparency by making it a requirement of funding in appropriate cases. In promising steps, the U.S. President and his administration declared 2023 the Year of Open Science,77 and the U.S. National Institutes of Health78 and the U.S. Department of Education79 both recently instituted data sharing policies for research they fund.

Finally, empirical legal research can take advantage of the larger movement in the social sciences, medicine, and many other fields, by leveraging the technology, training, and ideas flowing from those credibility revolutions. Free technologies like the Open Science Framework provide a place not just to store data, but to collaborate, establish version control, preregister, and store video stimuli. Other examples include tools like Github (a data and code repository), AsPredicted (a general study registry), Declare Design (a tool for creating a preregistration), and the American Economic Association’s registry for randomized controlled trials. Straightforward guides to data staring, preregistering, and many other transparency and credibility-related activities are now available.80 At least one guide specific to some empirical legal research methodologies is also available, and we hope more are on the way.81 With these tools at their fingertips – and as a field whose data and results are often of great public importance – there is little reason researchers in the field of empirical legal research should not become leaders in the move towards transparency and credibility.

Comments on this article Comments (0)

Version 2
VERSION 2 PUBLISHED 08 Feb 2023
Comment
Author details Author details
Competing interests
Grant information
Copyright
Download
 
Export To
metrics
Views Downloads
F1000Research - -
PubMed Central
Data from PMC are received and updated monthly.
- -
Citations
CITE
how to cite this article
Chin J, Zeiler K, Dilevski N et al. The transparency of quantitative empirical legal research published in highly ranked law journals (2018–2020): an observational study [version 2; peer review: 3 approved]. F1000Research 2024, 12:144 (https://doi.org/10.12688/f1000research.127563.2)
NOTE: If applicable, it is important to ensure the information in square brackets after the title is included in all citations of this article.
track
receive updates on this article
Track an article to receive email alerts on any updates to this article.

Open Peer Review

Current Reviewer Status: ?
Key to Reviewer Statuses VIEW
ApprovedThe paper is scientifically sound in its current form and only minor, if any, improvements are suggested
Approved with reservations A number of small changes, sometimes more significant revisions are required to address specific details and improve the papers academic merit.
Not approvedFundamental flaws in the paper seriously undermine the findings and conclusions
Version 1
VERSION 1
PUBLISHED 08 Feb 2023
Views
13
Cite
Reviewer Report 12 Oct 2023
Rob Heirene, School of Psychology, University of Plymouth, Plymouth, UK 
Approved
VIEWS 13
Many thanks for the invite to review this interesting manuscript.

Article summary: The article reports a meta-scientific review of 300 empirical articles published in the legal field in 2019-2020). The authors aimed to determine the level of ... Continue reading
CITE
CITE
HOW TO CITE THIS REPORT
Heirene R. Reviewer Report For: The transparency of quantitative empirical legal research published in highly ranked law journals (2018–2020): an observational study [version 2; peer review: 3 approved]. F1000Research 2024, 12:144 (https://doi.org/10.5256/f1000research.140083.r185631)
NOTE: it is important to ensure the information in square brackets after the title is included in all citations of this article.
  • Author Response 17 Jan 2024
    Jason Chin, College of Law, Australian National University, Canberra, Australia
    17 Jan 2024
    Author Response
    Thank you Rob for the feedback and helpful code!

    We were waiting for the final review to begin revisions. We expect to post them in mid-January 2024. - Jason ... Continue reading
COMMENTS ON THIS REPORT
  • Author Response 17 Jan 2024
    Jason Chin, College of Law, Australian National University, Canberra, Australia
    17 Jan 2024
    Author Response
    Thank you Rob for the feedback and helpful code!

    We were waiting for the final review to begin revisions. We expect to post them in mid-January 2024. - Jason ... Continue reading
Views
9
Cite
Reviewer Report 17 Aug 2023
Stefanie Mueller, Leibniz Institute for Psychology (ZPID), Trier, Germany 
Approved
VIEWS 9
The manuscript describes the adoption of transparency practices for studies in the field of legal research. The studies were published in either faculty-edited journals or in student-edited journals. Student-edited journals were selected based on their Washington & Lee ranking (which ... Continue reading
CITE
CITE
HOW TO CITE THIS REPORT
Mueller S. Reviewer Report For: The transparency of quantitative empirical legal research published in highly ranked law journals (2018–2020): an observational study [version 2; peer review: 3 approved]. F1000Research 2024, 12:144 (https://doi.org/10.5256/f1000research.140083.r185632)
NOTE: it is important to ensure the information in square brackets after the title is included in all citations of this article.
  • Author Response 17 Jan 2024
    Jason Chin, College of Law, Australian National University, Canberra, Australia
    17 Jan 2024
    Author Response
    Thank you Stefanie for these helpful comments. 

    We were waiting for the final review to begin revisions. We expect to post them in mid-January 2024. - Jason et al
    Competing Interests: None other than being the authors.
COMMENTS ON THIS REPORT
  • Author Response 17 Jan 2024
    Jason Chin, College of Law, Australian National University, Canberra, Australia
    17 Jan 2024
    Author Response
    Thank you Stefanie for these helpful comments. 

    We were waiting for the final review to begin revisions. We expect to post them in mid-January 2024. - Jason et al
    Competing Interests: None other than being the authors.
Views
34
Cite
Reviewer Report 27 Feb 2023
Antica Culina, Netherlands Institute of Ecology, NIOO-KNAW, Wageningen, The Netherlands 
Approved
VIEWS 34
The MS uses articles published in the field of empirical legal research to study the occurrence of several practices related to transparent research and credibility revolution – such as preregistration, open data and protocols, or open access to publication. These ... Continue reading
CITE
CITE
HOW TO CITE THIS REPORT
Culina A. Reviewer Report For: The transparency of quantitative empirical legal research published in highly ranked law journals (2018–2020): an observational study [version 2; peer review: 3 approved]. F1000Research 2024, 12:144 (https://doi.org/10.5256/f1000research.140083.r162735)
NOTE: it is important to ensure the information in square brackets after the title is included in all citations of this article.
  • Author Response 17 Jan 2024
    Jason Chin, College of Law, Australian National University, Canberra, Australia
    17 Jan 2024
    Author Response
    Thanks so much Antica - we were waiting for the final review to begin revisions. We expect to post them in mid-January 2024. - Jason et al
    Competing Interests: No competing interests other than the obvious one of being the authors.
COMMENTS ON THIS REPORT
  • Author Response 17 Jan 2024
    Jason Chin, College of Law, Australian National University, Canberra, Australia
    17 Jan 2024
    Author Response
    Thanks so much Antica - we were waiting for the final review to begin revisions. We expect to post them in mid-January 2024. - Jason et al
    Competing Interests: No competing interests other than the obvious one of being the authors.

Comments on this article Comments (0)

Version 2
VERSION 2 PUBLISHED 08 Feb 2023
Comment
Alongside their report, reviewers assign a status to the article:
Approved - the paper is scientifically sound in its current form and only minor, if any, improvements are suggested
Approved with reservations - A number of small changes, sometimes more significant revisions are required to address specific details and improve the papers academic merit.
Not approved - fundamental flaws in the paper seriously undermine the findings and conclusions
Sign In
If you've forgotten your password, please enter your email address below and we'll send you instructions on how to reset your password.

The email address should be the one you originally registered with F1000.

Email address not valid, please try again

You registered with F1000 via Google, so we cannot reset your password.

To sign in, please click here.

If you still need help with your Google account password, please click here.

You registered with F1000 via Facebook, so we cannot reset your password.

To sign in, please click here.

If you still need help with your Facebook account password, please click here.

Code not correct, please try again
Email us for further assistance.
Server error, please try again.