SOTAVerified

Towards Task-Agnostic Privacy- and Utility-Preserving Models

2021-09-01RANLP 2021Unverified0· sign in to hype

Yaroslav Emelyanov

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Modern deep learning models for natural language processing rely heavily on large amounts of annotated texts. However, obtaining such texts may be difficult when they contain personal or confidential information, for example, in health or legal domains. In this work, we propose a method of de-identifying free-form text documents by carefully redacting sensitive data in them. We show that our method preserves data utility for text classification, sequence labeling and question answering tasks.

Tasks

Reproductions