LLM Watermarking Using Mixtures and Statistical-to-Computational Gaps
2025-05-02Unverified0· sign in to hype
Pedro Abdalla, Roman Vershynin
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Given a text, can we determine whether it was generated by a large language model (LLM) or by a human? A widely studied approach to this problem is watermarking. We propose an undetectable and elementary watermarking scheme in the closed setting. Also, in the harder open setting, where the adversary has access to most of the model, we propose an unremovable watermarking scheme.