TextSeal: A Localized LLM Watermark for Provenance & Distillation Protection
Researchers have developed TextSeal, a novel watermarking technique for large language models designed to protect against unauthorized use and distillation. This method utilizes dual-key generation and entropy-weighted scoring for robust detection, even in mixed human-AI content. TextSeal maintains output diversity and does not introduce inference overhead, outperforming existing baselines while preserving downstream task performance and human-perceived quality. AI
IMPACT Introduces a new method to track and protect LLM outputs, potentially impacting model provenance and preventing unauthorized derivative works.