Hostname: page-component-6bb9c88b65-g7ldn Total loading time: 0 Render date: 2025-07-22T06:03:03.367Z Has data issue: false hasContentIssue false

Evaluation of semi-automated record screening methods for systematic reviews of prognosis studies and intervention studies

Published online by Cambridge University Press:  22 July 2025

Isa Spiero*
Affiliation:
Julius Center for Health Sciences and Primary Care, https://ror.org/0575yy874 University Medical Center Utrecht, Utrecht University , Utrecht, the Netherlands
Artuur M. Leeuwenberg
Affiliation:
Julius Center for Health Sciences and Primary Care, https://ror.org/0575yy874 University Medical Center Utrecht, Utrecht University , Utrecht, the Netherlands
Karel G. M. Moons
Affiliation:
Julius Center for Health Sciences and Primary Care, https://ror.org/0575yy874 University Medical Center Utrecht, Utrecht University , Utrecht, the Netherlands
Lotty Hooft
Affiliation:
Julius Center for Health Sciences and Primary Care, https://ror.org/0575yy874 University Medical Center Utrecht, Utrecht University , Utrecht, the Netherlands Cochrane Netherlands, Julius Center for Health Sciences and Primary Care, https://ror.org/0575yy874 University Medical Center Utrecht, Utrecht University , Utrecht, the Netherlands
Johanna A. A. Damen
Affiliation:
Julius Center for Health Sciences and Primary Care, https://ror.org/0575yy874 University Medical Center Utrecht, Utrecht University , Utrecht, the Netherlands Cochrane Netherlands, Julius Center for Health Sciences and Primary Care, https://ror.org/0575yy874 University Medical Center Utrecht, Utrecht University , Utrecht, the Netherlands
*
Corresponding author: Isa Spiero; Email: i.spiero-2@umcutrecht.nl
Rights & Permissions [Opens in a new window]

Abstract

Systematic reviews (SRs) synthesize evidence through a rigorous, labor-intensive, and costly process. To accelerate the title–abstract screening phase of SRs, several artificial intelligence (AI)-based semi-automated screening tools have been developed to reduce workload by prioritizing relevant records. However, their performance is primarily evaluated for SRs of intervention studies, which generally have well-structured abstracts. Here, we evaluate whether screening tool performance is equally effective for SRs of prognosis studies that have larger heterogeneity between abstracts. We conducted retrospective simulations on prognosis and intervention reviews using a screening tool (ASReview). We also evaluated the effects of review scope (i.e., breadth of the research question), number of (relevant) records, and modeling methods within the tool. Performance was assessed in terms of recall (i.e., sensitivity), precision at 95% recall (i.e., positive predictive value at 95% recall), and workload reduction (work saved over sampling at 95% recall [WSS@95%]). The WSS@95% was slightly worse for prognosis reviews (range: 0.324–0.597) than for intervention reviews (range: 0.613–0.895). The precision was higher for prognosis (range: 0.115–0.400) compared to intervention reviews (range: 0.024–0.057). These differences were primarily due to the larger number of relevant records in the prognosis reviews. The modeling methods and the scope of the prognosis review did not significantly impact tool performance. We conclude that the larger abstract heterogeneity of prognosis studies does not substantially affect the effectiveness of screening tools for SRs of prognosis. Further evaluation studies including a standardized evaluation framework are needed to enable prospective decisions on the reliable use of screening tools.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of The Society for Research Synthesis Methodology

Highlights

What is already known?

  • The conduct of SRs can be accelerated by AI-based semi-automated screening tools in which active learning and machine learning are combined to prioritize potentially relevant abstracts.

  • Previous studies using such tools have already shown substantial workload reductions in title–abstract screening. However, these tools are most often tested for reviews of intervention studies.

What is new?

  • We evaluated the performance of a screening tool for SRs of prognosis studies as the larger heterogeneity between abstracts of prognosis studies could negatively affect the tool performance.

  • The performance of the tool was only slightly decreased for prognosis reviews compared to intervention reviews. The scope of the review research question and modeling methods within the tool only slightly affected the screening tool performance, while the effect of the number of (relevant) records in the dataset was strongly related to tool performance.

Potential impact for RSM readers

  • With this study, we demonstrate that screening tools are almost equally effective for prognosis reviews as compared to intervention reviews.

  • We also gained insight into some additional characteristics of reviews and/or tool settings that affect tool performance. These insights in the context in which tools may perform well could eventually enable users to prospectively decide the suitability of tool use for a given review.

1 Introduction

Systematic reviews (SRs) and clinical guidelines form an essential part of evidence-based medicine by presenting thorough literature searches to collect and summarize all relevant primary studies to answer a given research question. It is essential to maximize the sensitivity of the literature search to reduce the amount of bias in the conclusions of the review or guideline. Hereto, a broad literature search in electronic databases is typically applied, frequently resulting in large numbers of records that need to be screened for relevance, first by title–abstract screening followed by full text screening. Conducted manually, this process is error prone, labor-intensive, and time-consuming, with the average review taking 1–2 years until completion.Reference Borah, Brown, Capers and Kaiser1 Moreover, as new evidence emerges faster by the increasing number of studies that are published each year,Reference Bornmann, Haunschild and Mutz2 a SR or guideline may already be outdated by the time it is published.Reference Beller, Chen, Wang and Glasziou3

After one of the first attempts to accelerate the screening process with automated classification of titles and abstracts by Cohen et al. Reference Cohen, Hersh, Peterson and Yen4 many artificial intelligence (AI)-based (semi-)automated screening tools and algorithms have been developed.Reference O’Mara-Eves, Thomas, McNaught, Miwa and Ananiadou5 These screening tools combine feature extraction models that draw relevant features from the titles and abstracts with classification models that subsequently use these features to predict probabilities of relevance for each study. In practice, such tools are implemented either fully automated by providing the user with a binary outcome label (i.e., “relevant” or “irrelevant”) for each of the records or, most often, semi-automated by not dichotomizing but rather ranking or prioritizing the records based on the predicted probability of relevance.Reference Hamel6 In the latter, records are ranked according to the predicted relevance, and researchers screen the records in that order. The screening tool is iteratively retrained by the decisions that the researcher makes, and the records are reranked accordingly. This is also called “active learning” and can be divided into certainty based (in which the researcher is first presented with the records that the algorithm is most certain about) and uncertainty sampling (in which the algorithm presents the records that it is least certain about).Reference Marshall and Wallace7 By using this procedure, semi-automated screening tools can have one of several functions such as (1) increasing rate of screening and/or improving workflow by prioritizing relevant records, (2) functioning as a second screener, and (3) reducing the number of records to screen by excluding records predicted to be irrelevant.Reference O’Mara-Eves, Thomas, McNaught, Miwa and Ananiadou5 Currently, most tools are semi-automatic and mainly function as a second screener or as improvement of workflow by prioritization during manual screening.Reference Marshall and Wallace7

Performance evaluations of semi-automated screening tools have shown an average workload reduction of about 50%, but this varies greatly across contexts.Reference Hamel6, Reference Tsou, Treadwell, Erinoff and Schoelles8 This heterogeneity in tool performance is presumably related to factors such as the tool and its modeling methods,Reference Adeva, Atxa, Carrillo and Zengotitabengoa9 number of records to be screened for the review,Reference Hamel6, Reference Feng10 number of labeled records used for training the models within the tool,Reference Tsou, Treadwell, Erinoff and Schoelles8, Reference Hopewell11 complexity/breadth of the research question,Reference Valizadeh12, Reference Damen, Moons, van Smeden and Hooft13 abstract structure and coherence,Reference Burgard and Bittermann14 number of relevant records,Reference Tsou, Treadwell, Erinoff and Schoelles8, Reference Van De Schoot15 review type (e.g., intervention, prognosis, or diagnosis), and/or medical topic .Reference Van De Schoot15, Reference Muthu16 However, the validation of a semi-automated screening tool is often based on performance evaluations with only one or just a few SRs,Reference Feng10 and each evaluation study uses different methods as there is no standardized way for evaluation.Reference O’Mara-Eves, Thomas, McNaught, Miwa and Ananiadou5 Therefore, it is difficult to make valid comparisons across evaluation studies to assess the effect of aforementioned factors on tool performance and to subsequently decide prospectively whether tool use is suitable in a given setting.

Within the medical field, tools have mostly been validated on (therapeutic or preventive) intervention reviews, especially concerning randomized clinical trials (RCTs). For RCTs, there is typically better adherence to strict abstract reporting guidelines,Reference Hopewell11 as compared to, for example, prognosis and diagnosis studies. Therefore, poor generalizability of performance to these types of studies is expected.Reference Khalil, Ameen and Zarnegar17 For diagnostic test accuracy reviews, for example, it was already suspected that the unreliable identification of eligible records by semi-automated screening was caused by the inconsistent terminology used in the abstracts of these studies.Reference Valizadeh12 Similarly, reports of prognosis studies are even more variable, and this heterogeneity between abstracts can even further increase for prognosis reviews that have topics with broader scopes (i.e., a broader research question to be answered by the review). For such broad scopes, a broader search strategy is typically applied, resulting in a larger number and a larger variability in retrieved records. In such cases, the algorithms of the tools may have more difficulty to find the features needed to predict a probability of relevance. At the same time, prognosis reviews generally have more records to be screened based on the search strategy,Reference Damen, Moons, van Smeden and Hooft13 and the potential workload reduction that can be obtained by using a semi-automated tool for these reviews is thereby higher.

We evaluated whether the performance of title–abstract screening methods is equally sufficient for SRs of prognosis studies compared to intervention studies. We illustrate this by using a currently available screening tool as a case study. Additionally, we investigated whether factors such as the scope of the review, the number of (relevant) records of the review, and the modeling methods used in the tool contribute to performance differences between these review types.

2 Methods

2.1 Title–abstract screening tool

From the currently available AI-based semi-automated screening tools for title–abstract screening, the ASReview toolReference Van De Schoot15 (hereafter called the screening tool) was chosen to compare the performance of semi-automated ranking algorithms when applied to SRs of prognosis and intervention studies. As a free, open-source tool based on Python, it allows researchers to flexibly conduct large amounts of simulations on screening tool use in contrast to other tools that are not open source. The tool has been evaluated for previous reviews including reviews within the medical domain.Reference Muthu16 Several options are offered by the tool to perform simulation studies on labeled review datasets, with different feature extraction, classification, and query models among others to be chosen by the user. The classification models include methods such as support vector machines (SVMs) and naive Bayes (NB) that are also incorporated in other commonly used screening tools (such as AbstrackR,Reference Wallace18 DistillerAI,19 EPPI-Reviewer,20 and RayyanReference Ouzzani, Hammady, Fedorowicz and Elmagarmid21) and could thereby give an indication of the performance of the algorithms within these tools as well. By using a labeled dataset and the chosen modeling methods, the screening tool conducts active learning in which the models are iteratively retrained to predict probabilities of relevance based on the titles and abstracts of the records and uses these probabilities to rank and prioritize the records to the user accordingly (Figure 1).

Figure 1 Overview of the active learning simulations that were conducted to evaluate the performance of a title–abstract screening tool. The scope breadth of the datasets, the modeling methods, and the number of (relevant) records were varied.

Table 1 Description of the review datasets used for evaluating the performance of a semi-automated title–abstract screening tool

a–c From broad to moderate to narrow review scope (i.e., the breadth of the research question of the review).

2.2 Datasets

To evaluate the performance of the screening tool, retrospective simulations were performed using a convenience sample of manually screened datasets of previously published prognosis (n = 6) and intervention reviews (n = 6) (Table 1). The reviews consist of a range of different topics and the number of records to be screened. The intervention review datasets are publicly available as being part of the Conference and Labs of the Evaluation Forum 2019 (CLEF 2019) challenge,Reference Kanoulas35 and for the current study, a selection was made of the intervention reviews with the largest number of records (ranging from 2,815 to 12,319 records) to match (and reduce the potential effect of) the larger size of prognosis reviews (ranging from 777 to 10,664 records). Since the intervention review by Kahn et al. (2018) did not have enough relevant records to run simulations with default settings (as described under “Simulations”), the results for this review were not included in the main analysis but can be found in the Supplementary Material instead (Part IV: Figure A3 and Table A8 in the Supplementary Material).

Each of the datasets consists of records with manually labeled relevant/irrelevant labels based on title and abstract screening, which were used as the reference standard. The percentages of the manually included relevant records were generally low for the six intervention reviews (ranging from 0.3 to 2.1% of all screened records), while for prognosis reviews, the percentages were much higher (ranging from 7.1 to 26.6%). In addition to the division into review type (prognosis vs intervention), a further subdivision based on the scope of the review (i.e., breadth of the research question) was made for prognosis reviews to further examine the effect of abstract heterogeneity on screening tool performance. The scopes of the prognosis reviews could be roughly divided into: (A) broad scope reviews that had no restrictions for the population, outcome, or prediction model (prognosis reviews 1 and 3), (B) moderate scope reviews that had restrictions for population and outcome but not for prediction model (prognosis reviews 4 and 5), and (C) narrow scope reviews that had restrictions for the population, outcome, and prediction model (prognosis reviews 2 and 6).

2.3 Simulations

The title–abstract screening process was simulated with the datasets of these 12 reviews. To initiate a simulation, a predefined number of 10 relevant and 10 irrelevant records was randomly sampled from the review dataset by the tool. This sample functioned as the training dataset for the model to initiate the record ranking based on the probability of inclusion. For the ranking, the trained model was applied by the tool to the remaining records, and for each record, the probability of relevance was computed. The records were then ranked according to their relevance probability from highest to lowest (using the default query method of “maximum probability”) and simulated as being presented to the tool user, by starting with the record with the highest probability and using the relevant/irrelevant label of the respective record as the supposed decision by the user. After having presented the predefined number of 10 highest-ranking records, the model was updated while taking these 10 records and their labels into account. Then, the process of presenting the highest-ranking records, retrieving labels, and model updating was iterated until all records had passed. Since the initial training set (of 10 relevant and 10 irrelevant records) was sampled randomly, the entire process was repeated 200 times, in which each time, a new set of relevant and irrelevant records was sampled. Furthermore, since the number of relevant records differed per dataset, some datasets may have suffered from class imbalance (i.e., few relevant records compared to irrelevant records) more than others. In the screening tool, we set the balancing strategy that deals with this class imbalance to the tool’s default strategy (double balance).

Several variations within the simulations were introduced to test their effects. First, we used a sample of prognosis reviews consisting of different scope breadths, as described under Datasets. Broader scopes result in a larger variation in retrieved studies due to the broader search strategy, and therefore, more heterogeneity between titles and abstracts was expected. This can, in turn, influence the performance of the screening tool in distinguishing between relevant and irrelevant records.

Second, we varied the modeling methods within the tool. These modeling methods consist of algorithms which combine a feature extraction model with a classification model. A feature extraction model extracts features from the text, such as how often words occur in the abstracts. These features can then be used by a classification model that learns, based on these features, which abstracts are relevant and irrelevant and can then apply this knowledge to make predictions for unseen abstracts (e.g., it may learn that the occurrence of “high blood pressure” may be indicative that the abstract is likely relevant in a review on antihypertensive medication). For each of the datasets, we tested two different feature extraction models (term frequency-inverse document frequency [TF-IDF] and sentence Bidirectional Encoder Representations from Transformers [sBERT]) and three different classification models (NB, logistic regression, and SVM), leading to five different feature extraction and classification model combinations (as sBERT is not compatible with NB):

  1. (1) TF-IDF with NB (screening tool default),

  2. (2) TF-IDF with SVM,

  3. (3) TF-IDF with logistic regression,

  4. (4) sBERT with SVM,

  5. (5) sBERT with logistic regression.

Third, we varied the number of records in the review datasets to compare intervention reviews to prognosis reviews while ruling out the influence of review size and/or number of relevant records. The datasets were manually adapted to consist of a random sample of 500, 1,000, and 2,000 records, of which 50 were relevant. Each random sample from the original datasets was drawn five times. This was conducted only for the eight review datasets that allowed this in terms of available relevant records (i.e., intervention reviews 1, 2, 4, and 6 and prognosis reviews 3, 4, 5, and 6). Thus, an additional number of 24 manually adapted review datasets were simulated 200 times each, but here only with the tool’s default modeling settings (TF-IDF with NB).

2.4 Performance evaluation

A 2x2 cross-tabulation was used in which the reference standard of manual relevant/irrelevant labels was compared to predictions of the screening tool at a given percentage of records screened. The following sections describe the choice of metrics to quantify and compare the performance of the screening tool. The formulas for calculations and corresponding definitions of each metric can be found in the Supplementary Material (Part I).

The recall (or sensitivity) at a given number of records screened indicates the proportion of correctly classified positives by the tool. For SRs in the medical field, a recall of at least 95% is desired to prevent the introduction of bias in the results. However, at each given number of records screened, the number of false negatives is by default more likely to be lower when less relevant records are present in the dataset, and therefore, the recall is subsequently also more likely to be higher in such datasets compared to datasets that contain more relevant records. Thus, given this dependence of the recall on the number of relevant records that are present in the dataset, we additionally provided the maximum achievable recall at increasing numbers of records screened. This maximum value is dependent on the number of relevant records in the dataset; for example, if you have a tool with perfect performance (only returns true relevant records) and one screens 100 out of 1,000 records (of which 150 are relevant), the maximum achievable recall would be 100 (TP)/(100 (TP) + 50 (FN)) = 0.67, while if one screens 100 out of 1,000 records (of which 100 are relevant) the maximum achievable recall would be 100 (TP)/(100 (TP) + 0 (FN) = 1 already.

The precision (or positive predictive value) is defined by the proportion of correctly classified positives among the total number of classified positives and was calculated at the point that 95% recall was reached (precision@95%). The precision relates to the screening burden, and a higher value indicates that fewer irrelevant records need to be screened.

The Work Saved over Sampling (WSS) introduced by Cohen et al. Reference Cohen, Hersh, Peterson and Yen4 is defined as the proportion of records of the entire dataset that would not have to be screened after a given value of recall has been achieved and was also calculated at 95% recall (WSS@95%). Due to the dependence of the WSS on the number of relevant records, we additionally computed the normalized WSS (n-WSS@95%) introduced by Kusa et al. Reference Kusa, Lipani, Knoth and Hanbury36 for better comparison between the prognosis and intervention reviews as the latter have fewer relevant records.

The absolute workload reduction was defined by both the number of records not needed to screen (assuming the reviewer could stop screening at 95% recall) and by the corresponding amount of time saved in hours. To calculate the time saved, an average duration of 30s was considered as the time needed to screen one record, i.e., title and abstract.

3 Results

3.1 Screening tool performance in original datasets

Based on the original datasets, the intervention reviews (Figure 1) generally reached a recall of 95% earlier in the simulations of the screening process compared to prognosis reviews (Figure 2), that is, a smaller percentage of records needs to be screened for intervention reviews to find 95% of the relevant records. This is also apparent from the higher WSS@95% values that were achieved for intervention reviews (WSS@95% range: 0.613–0.895) compared to prognosis reviews (WSS@95% range: 0.324–0.597), and this difference in WSS between intervention and prognosis reviews remained after normalization of the WSS@95% (n-WSS@95% range 0.674–0.946 and 0.454–0.701, respectively) (Part II: Table A1 in the Supplementary Material; default TF-IDF with NB models). This corresponds to a workload reduction of 17.4–76.4 hours (2.087–9.137 records) for intervention reviews compared to a workload reduction of 3.6–54.4 hours (430–6.524 records) for prognosis reviews. However, the precision@95% was overall much lower for intervention reviews (precision range: 0.024–0.057) compared to prognosis reviews (precision range: 0.115–0.400) (Part II: Table A1 in the Supplementary Material; default TF-IDF with NB models).

Figure 2 Performance in terms of recall at increasing percentages of records screened for prognosis reviews.

Abbreviations: logistic: logistic regression; NB: Naive Bayes; SVM: support vector machine; sBERT: sentence Bidirectional Encoder Representations from Transformers; TF-IDF: term frequency-inverse document frequency.

3.2 Variations between prognosis review scopes

With regard to the effect of the scope of the prognosis reviews on the performance of the screening tool, there was no indication of increased performance for reviews with narrower scopes (prognosis reviews 6 and 2), compared to moderate scopes (prognosis reviews 4 and 5) or to broad scopes (prognosis reviews 1 and 3) based on the simulations of the original datasets (Figures 2 and 3).

Figure 3 Performance in terms of recall at increasing percentages of records screened for intervention reviews.

Abbreviations: logistic: Logistic Regression; NB: Naive Bayes; SVM: Support Vector Machine; sBERT: sentence Bidirectional Encoder Representations from Transformers; TF-IDF: Term Frequency-Inverse Document Frequency.

3.3 Variations in modeling methods within the screening

The simulations showed that the five modeling methods within the screening tool did not clearly differ in screening performance, except for a slight decreased performance when using sBERT as the feature extraction model in intervention reviews 2 and 4 and in prognosis review 2 compared to using TF-IDF (Figures 1 and 2). These and some slight differences between other models, if present, were usually found earlier in the screening process and would therefore not affect the WSS@95% or precision@95% (Part II: Table A1 in the Supplementary Material). Given this similarity in performance between the models, the remaining results were focused on the default models (TF-IDF with NB) only.

3.4 Variations between the number of (relevant) records

When the total number of records and the number of relevant records were equal between intervention and prognosis reviews after manual adaptation (500, 1,000, or 2,000 records with 50 relevant), the performance of the prognosis reviews still did not fully match the intervention reviews (Part II: Figures A2–A7 in the Supplementary Material). The WSS@95% was generally slightly larger for the intervention reviews (WSS@95% range: 0.446–0.865) compared to the prognosis reviews (WSS@95% range 0.355–0.738) and also after normalization (n-WSS@95% range: 0.513–0.939 and n-WSS@95% range: 0.410–0.801, respectively) (Part III: Table A2 in the Supplementary Material). The precision was also slightly higher for intervention reviews (precision@95% range: 0.041–0.5819) compared to prognosis reviews (precision@95% range: 0.034–0.291) (Part III: Table A2 in the Supplementary Material). In addition to this slight difference in performance between prognosis and intervention reviews, the specific data that was sampled in the adapted datasets (i.e., the fixed numbers of randomly sampled relevant and irrelevant records) also affected the performance of the screening tool (Part III: Figures A2 and A3 in the Supplementary Material).

4 Discussion

The performance of AI-based semi-automated title–abstract screening methods was compared between prognosis studies and intervention studies. We selected one of the available screening tools (ASReview) to illustrate the performance of semi-automated title–abstract screening. The amount of work saved, expressed as the (n-)WSS@95%, was slightly larger for intervention reviews, but the precision (i.e., positive predictive value) of the tool was lower compared to prognosis reviews. When the proportion of relevant records was manipulated by randomly sampling equal numbers of records for intervention and prognosis review datasets, the WSS@95% was still slightly larger for intervention reviews, and also, the precision was larger compared to prognosis reviews. Furthermore, there were only minor effects of modeling methods in either review type, and there was no clear effect of the scope (i.e., breadth of the research question) of the prognosis review on the performance of the tool.

As would be expected, the screening tool performed slightly worse for prognosis reviews compared to intervention reviews, but this difference was not substantial. It is plausible that this lower performance was due to the larger heterogeneity in abstracts of prognosis studies, causing the feature extraction models to have more difficulty in learning the appropriate features to subsequently classify the abstracts as relevant or irrelevant. Accordingly, the percentage of work saved was slightly higher for intervention reviews. However, the absolute number of records that needs to be screened is, in practice, generally higher for prognosis reviews. Therefore, this implies that the absolute number of records that do not need to be screened, thus the absolute amount of work saved, could in practice be much larger for prognosis reviews. Future work could further evaluate the potential impact of abstract heterogeneity on screening tool performance by incorporating reviews from other domains that may be more affected by this issue.

It should be noted that a considerable part of the initial difference in performance between the original datasets of prognosis and intervention reviews could be explained by the lower number of actually relevant records in the intervention review datasets. A low number of relevant records impacts the maximum achievable values of the performance metrics (i.e., these values are increased for WSS and recall and decreased for precision), which was compensated for in this study by our selection of intervention and prognosis reviews based on similar number of records and by the additional analysis in which equal numbers of (included) records were selected from the reviews for simulations. Consequently, this highlights that evaluating performance on datasets with limited variation in numbers of relevant records and with limited reporting of metrics may give misleading impressions on the general performance of a screening tool. The same applies to the size of the dataset; the maximum achievable value for WSS is inherently related to the total number of records (under a constant number of relevant records). Nevertheless, these aspects are hardly ever addressed in comparative evaluation or validation studies of semi-automated title–abstract screening tools but should be considered along with the need for standardized evaluation and reporting.

Interestingly, the performance of the tool did not differ between the size of the scope of the prognosis reviews. Nevertheless, there were only a few reviews for each scope breadth, and the true relation between scope breadth and coherence of the relevant abstracts was not further assessed in this study, but could influence the model performance.Reference Burgard and Bittermann14 Furthermore, similar to the results of the comparative review on screening tool performance by Feng et al. Reference Feng10 there were only minor differences in performance between the modeling methods. These differences mainly involved the sBERT feature extraction model which not only took longer to train due to its complexity, but also performed similar to even slightly worse compared to the simple method of TF-IDF as has also previously been suggested.Reference Nowak37

Even though this study indicates that the performance of semi-automated screening tools is not substantially different between SRs of prognosis and intervention studies, there are still some issues that need to be overcome before the implementation of semi-automated tools in either review type. First, there is still a lack of trust by the SR community in the adoption of these toolsReference O’Connor, Tsafnat and Thomas38 which could be related to the lack of standardized evaluation and (external) validation of tools.Reference O’Connor, Glasziou and Taylor39 Without standardized ways of evaluation of tool performance, it becomes difficult to compare tools and to draw conclusions on the generalizability of performance to other contexts. Although there is a lack of validation studies, the number of studies on developing new tools or algorithms is growing,Reference O’Mara-Eves, Thomas, McNaught, Miwa and Ananiadou5 making end users lose track of what tools they can use for their specific context.

Furthermore, there is still no consensus reached on what the appropriate stopping criterion for semi-automated screening should be in order to balance the trade-off between screening costs and sensitivity of the classification, including how this can prospectively be defined with sufficient certainty. As a result, such stopping criteria have not been incorporated in most of the currently available ready-to-use tools.Reference Nowak37 Stopping criteria that have been proposed include time-based or pragmatic approaches (i.e., stopping after a certain time or percentage screened), heuristic or data-driven approaches (i.e., stopping after encountering a predefined number of consecutive irrelevant records), and sampling approaches (i.e., estimating the number of relevant records based on screening a sample of the records). More statistical methods have also been proposed which use more advanced methods to prospectively estimate the recall including its certainty.Reference Di Nunzio, Pasi, Piwowarski, Azzopardi and Hanbury40, Reference Callaghan and Müller-Hansen41

Limitations of the current study include the reference standard of manual screening that was used in the simulations. This standard is likely to contain errors due to screening fatigue and screeners’ over-inclusiveness.Reference O’Mara-Eves, Thomas, McNaught, Miwa and Ananiadou5 Since this effect would be the same for the intervention and prognosis datasets, it could be assumed to not have affected the comparison between the two review types. The datasets consisted of a convenience sample and are thereby no true representation of all reviews of either type which could possibly have led to selection bias. The records eligible for full text screening were used as the reference standard as opposed to records that were final inclusions in the review, in order to have the simulations resemble title–abstract screening in practice. As such, we have no clear indication of the impact of the missed records during title–abstract screening on missing final inclusions in the review and subsequently on potential bias in the results.

5 Conclusion

In this study, we evaluated the performance of an AI-based screening tool for semi-automated title–abstract screening in the development of SRs for both intervention and prognosis studies. While intervention reviews have been extensively studied in the context of screening tool evaluations, prognosis reviews have not yet been thoroughly assessed. Despite the anticipated lower performance for prognosis reviews due to the larger heterogeneity of abstracts, the screening tool still demonstrated significant amounts of workload reduction for prognosis reviews that were only slightly lower compared to intervention reviews. Furthermore, we found no significant effects of the review’s scope (i.e., breadth of the research question of the review) or the specific modeling methods applied within the tool on its performance for either type of review. Instead, factors that did significantly affect screening tool performance were related to the number of records retrieved and the proportion of relevant records within the review. Given these findings, further evaluation studies should focus on developing methods to predict the tool’s performance based on the number of records retrieved and the expected proportion of relevant records. Nevertheless, the implementation of a screening tool in which active learning is combined with machine learning algorithms has been shown to be sufficiently applicable for semi-automated screening in both intervention and prognosis SRs.

Author contributions

Conceptualization: I.S., A.M.L., K.G.M., L.H., J.A.A.D.; Formal analysis: I.S.; Investigation: I.S.; Methodology: I.S., A.M.L., J.A.A.D.; Supervision: K.G.M., L.H.; Writing—original draft: I.S.; Writing—review and editing: I.S., A.M.L., K.G.M., L.H., J.A.A.D.

Competing interest statement

The authors declare that no competing interests exist.

Data availability statement

The code and data of our study are available at https://github.com/isa-sp/SR_semi-automated_screening. The data were partially derived from the publicly available data from the CLEF 2019 eHealth task at https://github.com/CLEF-TAR/tar/.

Funding statement

The authors declare that no specific funding has been received for this article.

Supplementary material

To view supplementary material for this article, please visit http://doi.org/10.1017/rsm.2025.10025.

Footnotes

This article was awarded Open Data and Open Materials badges for transparent practices. See the Data availability statement for details.

References

Borah, R, Brown, AW, Capers, PL, Kaiser, KA. Analysis of the time and workers needed to conduct systematic reviews of medical interventions using data from the PROSPERO registry. BMJ Open. 2017;7(2):e012545. https://doi.org/10.1136/bmjopen-2016-012545.CrossRefGoogle ScholarPubMed
Bornmann, L, Haunschild, R, Mutz, R. Growth rates of modern science: A latent piecewise growth curve approach to model publication numbers from established and new literature databases. Humanit Soc Sci Commun. 2021;8(1):Article no. 1. https://doi.org/10.1057/s41599-021-00903-w.CrossRefGoogle Scholar
Beller, EM, Chen, JK-H, Wang, UL-H, Glasziou, PP. Are systematic reviews up-to-date at the time of publication? Syst Rev. 2013;2(1):36. https://doi.org/10.1186/2046-4053-2-36.CrossRefGoogle ScholarPubMed
Cohen, AM, Hersh, WR, Peterson, K, Yen, PY. Reducing workload in systematic review preparation using automated citation classification. J Am Med Inform Assoc. 2006;13(2):206219. https://doi.org/10.1197/jamia.M1929.CrossRefGoogle ScholarPubMed
O’Mara-Eves, A, Thomas, J, McNaught, J, Miwa, M, Ananiadou, S. Using text mining for study identification in systematic reviews: A systematic review of current approaches. Syst Rev. 2015;4(1):5. https://doi.org/10.1186/2046-4053-4-5.CrossRefGoogle Scholar
Hamel, C, et al. Guidance for using artificial intelligence for title and abstract screening while conducting knowledge syntheses. BMC Med Res Methodol. 2021;21:112.CrossRefGoogle ScholarPubMed
Marshall, IJ, Wallace, BC. Toward systematic review automation: A practical guide to using machine learning tools in research synthesis. Syst Rev. 2019;8:10. https://doi.org/10.1186/s13643-019-1074-9.CrossRefGoogle ScholarPubMed
Tsou, AY, Treadwell, JR, Erinoff, E, Schoelles, K. Machine learning for screening prioritization in systematic reviews: Comparative performance of Abstrackr and EPPI-reviewer. Syst Rev. 2020;9:114.CrossRefGoogle ScholarPubMed
Adeva, JG, Atxa, JP, Carrillo, MU, Zengotitabengoa, EA. Automatic text classification to support systematic reviews in medicine. Expert Syst Appl. 2014;41(4):14981508.10.1016/j.eswa.2013.08.047CrossRefGoogle Scholar
Feng, Y, et al. Automated medical literature screening using artificial intelligence: A systematic review and meta-analysis. J Am Med Inform Assoc. 2022;29(8):14251432.CrossRefGoogle ScholarPubMed
Hopewell, S, et al. CONSORT for reporting randomized controlled trials in journal and conference abstracts: Explanation and elaboration. PLoS Med. 2008;5(1):e20. https://doi.org/10.1371/journal.pmed.0050020.CrossRefGoogle Scholar
Valizadeh, A, et al. Abstract screening using the automated tool rayyan: Results of effectiveness in three diagnostic test accuracy systematic reviews. BMC Med Res Methodol. 2022;22(1):115.CrossRefGoogle ScholarPubMed
Damen, JAA, Moons, KGM, van Smeden, M, Hooft, L. How to conduct a systematic review and meta-analysis of prognostic model studies. Clin Microbiol Infect. 2023;29(4):434440. https://doi.org/10.1016/j.cmi.2022.07.019.CrossRefGoogle ScholarPubMed
Burgard, T, Bittermann, A. Reducing literature screening workload with machine learning. Z Psychol. 2023;231(1):315. https://doi.org/10.1027/2151-2604/a000509.Google Scholar
Van De Schoot, R, et al. An open source machine learning framework for efficient and transparent systematic reviews. Nat Mach Intell. 2021;3(2):125133.10.1038/s42256-020-00287-7CrossRefGoogle Scholar
Muthu, S. The efficiency of machine learning-assisted platform for article screening in systematic reviews in orthopaedics. Int Orthop. 2023;47(2):551556. https://doi.org/10.1007/s00264-022-05672-y.CrossRefGoogle ScholarPubMed
Khalil, H, Ameen, D, Zarnegar, A. Tools to support the automation of systematic reviews: A scoping review. J Clin Epidemiol. 2022;144:2242. https://doi.org/10.1016/j.jclinepi.2021.12.005.CrossRefGoogle ScholarPubMed
Wallace, BC, et al. Deploying an interactive machine learning system in an evidence-based practice center: abstrackr. Proceedings of the 2nd ACM SIGHIT international health informatics symposium. 2012:819824. https://doi.org/10.1145/2110363.2110464.CrossRefGoogle Scholar
Eppi-Centre. EPPI-reviewer web. 2022. https://eppi.ioe.ac.uk/eppireviewer-web/home.Google Scholar
Ouzzani, M, Hammady, H, Fedorowicz, Z, Elmagarmid, A. Rayyan—A web and mobile app for systematic reviews. Syst Rev. 2016;5:110.CrossRefGoogle ScholarPubMed
Navarro, CLA, et al. Completeness of reporting of clinical prediction models developed using supervised machine learning: A systematic review. BMC Med Res Methodol. 2022;22(1):12. https://doi.org/10.1186/s12874-021-01469-6.CrossRefGoogle Scholar
Damen, JA, et al. Performance of the Framingham risk models and pooled cohort equations for predicting 10-year risk of cardiovascular disease: A systematic review and meta-analysis. BMC Med. 2019;17(1):109. https://doi.org/10.1186/s12916-019-1340-7.CrossRefGoogle ScholarPubMed
Heus, P, et al. Poor reporting of multivariable prediction model studies: Towards a targeted implementation strategy of the TRIPOD statement. BMC Med. 2018;16(1):120. https://doi.org/10.1186/s12916-018-1099-2.CrossRefGoogle ScholarPubMed
Pladet, LCA, et al. Prognostic models for mortality risk in patients requiring ECMO. Intensive Care Med. 2023;49(2):131141. https://doi.org/10.1007/s00134-022-06947-z.CrossRefGoogle ScholarPubMed
Takada, T, et al. Prognostic models for radiation-induced complications after radiotherapy in head and neck cancer patients. Cochrane Database Syst Rev. 2021;5:CD014745. https://doi.org/10.1002/14651858.CD014745.Google Scholar
Vernooij, LM, van Klei, WA, Moons, KG, et al. The comparative and added prognostic value of biomarkers to the revised cardiac risk index for preoperative prediction of major adverse cardiac events and all-cause mortality in patients who undergo noncardiac surgery. Cochrane Database Syst Rev. 2021;12:CD013139. https://www.cochranelibrary.com/cdsr/doi/10.1002/14651858.CD013139.pub2/abstract.Google Scholar
Arikpo, D, Edet, ES, Chibuzor, MT, Odey, F, Caldwell, DM. Educational interventions for improving primary caregiver complementary feeding practices for children aged 24 months and under. Cochrane Database Syst Rev. 2018;5(5):CD011768. https://www.cochranelibrary.com/cdsr/doi/10.1002/14651858.CD011768.pub2/abstract.Google ScholarPubMed
Chen, YJ, et al. First-line drugs inhibiting the renin angiotensin system versus other first-line antihypertensive drug classes for hypertension. Cochrane Database Syst Rev. 2018;11:CD008170. https://www.cochranelibrary.com/cdsr/doi/10.1002/14651858.CD008170.pub3/abstract.Google ScholarPubMed
Ijaz, S, Davies, P, Kessler, D, Lewis, G, Wiles, N. Psychological therapies for treatment-resistant depression in adults. Cochrane Database Syst Rev. 2018;5(5):CD010558. https://www.cochranelibrary.com/cdsr/doi/10.1002/14651858.CD010558.pub2/abstract.Google ScholarPubMed
Kahale, LA, et al. Anticoagulation for people with cancer and central venous catheters. Cochrane Database Syst Rev. 2018;6(6):CD006468. https://www.cochranelibrary.com/cdsr/doi/10.1002/14651858.CD006468.pub6/abstract.Google ScholarPubMed
Kaufman, J, et al. Face-to-face interventions for informing or educating parents about early childhood vaccination. Cochrane Database Syst Rev. 2018;5(5):CD010038. https://www.cochranelibrary.com/cdsr/doi/10.1002/14651858.CD010038.pub3/abstract.Google ScholarPubMed
Solomon, SD, Lindsley, K, Vedula, SS, Krzystolik, MG, Hawkins, BS. Anti-vascular endothelial growth factor for neovascular age-related macular degeneration. Cochrane Database Syst Rev. 2019;3(3):CD005139. https://doi.org/10.1002/14651858.CD005139.pub4.Google ScholarPubMed
Kahn, SR, et al. Interventions for implementation of thromboprophylaxis in hospitalized patients at risk for venous thromboembolism. Cochrane Database Syst Rev. 2018;4(4):CD008201. https://www.cochranelibrary.com/cdsr/doi/10.1002/14651858.CD008201.pub3/abstract.Google ScholarPubMed
Kanoulas, E, et al. CLEF 2019 technology assisted reviews in empirical medicine overview. CEUR workshop proceedings. 2019;2380:250. https://ceur-ws.org/Vol-2380/.Google Scholar
Kusa, W, Lipani, A, Knoth, P and Hanbury, A An analysis of work saved over sampling in the evaluation of automated citation screening in systematic literature reviews. Intell Syst Appl. 2023;18:200193. https://doi.org/10.1016/j.iswa.2023.200193.Google Scholar
Nowak, AJ. Artificial intelligence in evidence-based medicine. In: Lidströmer N, Ashrafian H, eds. Artificial Intelligence in Medicine. Springer, Cham; 2022: 255266. https://doi.org/10.1007/978-3-030-64573-1_432022.CrossRefGoogle Scholar
O’Connor, AM, Tsafnat, G, Thomas, J, et al. A question of trust: Can we build an evidence base to gain trust in systematic review automation technologies? Syst Rev. 2019;8(1):143. https://doi.org/10.1186/s13643-019-1062-0.CrossRefGoogle ScholarPubMed
O’Connor, AM, Glasziou, P, Taylor, M, et al. A focus on cross-purpose tools, automated recognition of study design in multiple disciplines, and evaluation of automation tools: A summary of significant discussions at the fourth meeting of the international collaboration for automation of systematic reviews (ICASR). Syst Rev. 2020;9(1):100. https://doi.org/10.1186/s13643-020-01351-4.CrossRefGoogle ScholarPubMed
Di Nunzio, GM. A study of an automatic stopping strategy for technologically assisted medical reviews. In: Pasi, G, Piwowarski, B, Azzopardi, L, Hanbury, A, eds. Advances in Information Retrieval, in Lecture Notes in Computer Science, 10772. Springer International Publishing; 2018:672677. https://doi.org/10.1007/978-3-319-76941-7_61.CrossRefGoogle Scholar
Callaghan, MW, Müller-Hansen, F. Statistical stopping criteria for automated screening in systematic reviews. Syst Rev. 2020;9(1):114.CrossRefGoogle ScholarPubMed
Figure 0

Figure 1 Overview of the active learning simulations that were conducted to evaluate the performance of a title–abstract screening tool. The scope breadth of the datasets, the modeling methods, and the number of (relevant) records were varied.

Figure 1

Table 1 Description of the review datasets used for evaluating the performance of a semi-automated title–abstract screening tool

Figure 2

Figure 2 Performance in terms of recall at increasing percentages of records screened for prognosis reviews.Abbreviations: logistic: logistic regression; NB: Naive Bayes; SVM: support vector machine; sBERT: sentence Bidirectional Encoder Representations from Transformers; TF-IDF: term frequency-inverse document frequency.

Figure 3

Figure 3 Performance in terms of recall at increasing percentages of records screened for intervention reviews.Abbreviations: logistic: Logistic Regression; NB: Naive Bayes; SVM: Support Vector Machine; sBERT: sentence Bidirectional Encoder Representations from Transformers; TF-IDF: Term Frequency-Inverse Document Frequency.

Supplementary material: File

Spiero et al. supplementary material

Spiero et al. supplementary material
Download Spiero et al. supplementary material(File)
File 2.1 MB