SOTAVerified

Human Perception in Natural Language Generation

2021-08-01ACL (GEM) 2021Unverified0· sign in to hype

Lorenzo De Mattei, Huiyuan Lai, Felice Dell’Orletta, Malvina Nissim

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We ask subjects whether they perceive as human-produced a bunch of texts, some of which are actually human-written, while others are automatically generated. We use this data to fine-tune a GPT-2 model to push it to generate more human-like texts, and observe that this fine-tuned model produces texts that are indeed perceived more human-like than the original model. Contextually, we show that our automatic evaluation strategy well correlates with human judgements. We also run a linguistic analysis to unveil the characteristics of human- vs machine-perceived language.

Tasks

Reproductions