SOTAVerified

Exploring Naive Approaches to Tell Apart LLMs Productions from Human-written Text

2024-03-05NLPIR '23: Proceedings of the 2023 7th International Conference on Natural Language Processing and Information Retrieval 2024Code Available0· sign in to hype

Oliver Giudice, Alessandro Maggi, Matteo Nardelli

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Powerful Large Language Models (large LMs or LLMs) such as BERT and GPT are making the task of detecting machine-generated text more and more prominent and crucial to minimize threats posed by text generation models misuse. Nonetheless, only a limited number of efforts exist so far, which can be classified into simple classifiers, zero-shot approaches, and fine-tuned LMs. These approaches usually rely on LMs whose discrimination accuracy decreases as the size difference in favor of the generator model increases (hence, a detector should always employ a LM with at least the same number of parameters of the source LM). Also, most of these approaches do not explicitly investigate whether the sentence syntactic structure can provide additional information that helps to build better detectors. All these considerations make the generalizing ability of detection methods into question. While generation techniques become more and more capable of producing human-like text, are the detection techniques capable of keeping up if not properly trained? In this paper, we evaluate the most effective (and reproducible) detection method available in the state of the art in order to figure out the limits in its robustness. We complement this analysis by discussing results obtained using a novel naive approach that demonstrably achieves comparable results in terms of robustness with respect to much more advanced and sophisticated state-of-the-art methods. Code with details on experiments are available at: https://github.com/bancaditalia/gen-text-detect.

Tasks

Reproductions