Abstract

Background
Conducting a systematic review requires great screening effort. Various tools have been proposed to speed up the process of screening thousands of titles and abstracts by engaging in active learning. In such tools, the reviewer interacts with machine learning software to identify relevant publications as early as possible. To gain a comprehensive understanding of active learning models for reducing workload in systematic reviews, the current study provides a methodical overview of such models. Active learning models were evaluated across four different classification techniques (naive Bayes, logistic regression, support vector machines, and random forest) and two different feature extraction strategies (TF-IDF and doc2vec). Moreover, models were evaluated across six systematic review datasets from various research areas to assess generalizability of active learning models across different research contexts.

Methods
Performance of the models were assessed by conducting simulations on six systematic review datasets. We defined desirable model performance as maximizing recall while minimizing the number of publications needed to screen. Model performance was evaluated by recall curves, WSS@95, RRF@10, and ATD.

Results
Within all datasets, the model performance exceeded screening at random order to a great degree.
The models reduced the number of publications needed to screen by 91.7% to 63.9%.

Conclusions
Active learning models for screening prioritization show great potential in reducing the workload in systematic reviews. Overall, the Naive Bayes + TF-IDF model performed the best.
Original languageEnglish
PublisherOSFPREPRINTS
Number of pages27
DOIs
Publication statusPublished - 16 Sept 2020

Fingerprint

Dive into the research topics of 'Active learning for screening prioritization in systematic reviews - A simulation study'. Together they form a unique fingerprint.

Cite this