Addressing the challenges of reconstructing systematic reviews datasets: a case study and a noisy label filter procedure.
Rutger NeelemanCathalijn H C LeenaarsMatthijs OudFelix WeijdemaRens van de SchootPublished in: Systematic reviews (2024)
Systematic reviews and meta-analyses typically require significant time and effort. Machine learning models have the potential to enhance screening efficiency in these processes. To effectively evaluate such models, fully labeled datasets-detailing all records screened by humans and their labeling decisions-are imperative. This paper presents the creation of a comprehensive dataset for a systematic review of treatments for Borderline Personality Disorder, as reported by Oud et al. (2018) for running a simulation study. The authors adhered to the PRISMA guidelines and published both the search query and the list of included records, but the complete dataset with all labels was not disclosed. We replicated their search and, facing the absence of initial screening data, introduced a Noisy Label Filter (NLF) procedure using active learning to validate noisy labels. Following the NLF application, no further relevant records were found. A simulation study employing the reconstructed dataset demonstrated that active learning could reduce screening time by 82.30% compared to random reading. The paper discusses potential causes for discrepancies, provides recommendations, and introduces a decision tree to assist in reconstructing datasets for the purpose of running simulation studies.
Keyphrases
- meta analyses
- systematic review
- machine learning
- randomized controlled trial
- borderline personality disorder
- rna seq
- minimally invasive
- clinical practice
- big data
- working memory
- electronic health record
- artificial intelligence
- human health
- decision making
- single cell
- risk assessment
- neural network
- positron emission tomography
- virtual reality
- pet imaging