Skip to yearly menu bar Skip to main content


Poster session B
in
Workshop: ICLR 2025 Workshop on GenAI Watermarking (WMARK)

Learning to watermark LLM-Generated Text Via Reinforcement Learning

Xiaojun Xu · Yuanshun Yao · Yang Liu


Abstract:

We study how to watermark LLM outputs, i.e. embedding algorithmically detectable signals into LLM-generated text to track misuse. Unlike the current mainstream methods that work with a fixed LLM, we expand the watermark design space by including the LLM tuning stage in the watermark pipeline. We propose a co-training framework based on reinforcement learning that iteratively (1) trains a detector to detect the generated watermarked text and (2) tunes the LLM to generate text easily detectable by the detector while keeping its normal utility. We empirically show that our watermarks are more accurate, robust, and adaptable (to new attacks) with no generation overhead. It also allows watermarked model open-sourcing. In addition, if used together with alignment, the extra overhead introduced is low -- we only need to train an extra reward model (i.e. our detector). We hope our work can bring more effort into studying a broader watermark design that is not limited to working with LLMs with unchanged model weights.

Chat is not available.