Towards Task-Agnostic Privacy- and Utility-Preserving Models
2021-09-01RANLP 2021Unverified0· sign in to hype
Yaroslav Emelyanov
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Modern deep learning models for natural language processing rely heavily on large amounts of annotated texts. However, obtaining such texts may be difficult when they contain personal or confidential information, for example, in health or legal domains. In this work, we propose a method of de-identifying free-form text documents by carefully redacting sensitive data in them. We show that our method preserves data utility for text classification, sequence labeling and question answering tasks.