SOTAVerified

Evaluating Pre-Trained Language Models on Multi-Document Summarization for Literature Reviews

2022-10-01sdp (COLING) 2022Unverified0· sign in to hype

Benjamin Yu

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Systematic literature reviews in the biomedical space are often expensive to conduct. Automation through machine learning and large language models could improve the accuracy and research outcomes from such reviews. In this study, we evaluate a pre-trained LongT5 model on the MSLR22: Multi-Document Summarization for Literature Reviews Shared Task datasets. We weren’t able to make any improvements on the dataset benchmark, but we do establish some evidence that current summarization metrics are insufficient in measuring summarization accuracy. A multi-document summarization web tool was also built to demonstrate the viability of summarization models for future investigators: https://ben-yu.github.io/summarizer

Tasks

Reproductions