Zhendong Zhao
2025
Watermarking with Low-Entropy POS-Guided Token Partitioning and Z-Score-Driven Dynamic Bias for Large Language Models
He Li
|
Xiaojun Chen
|
Zhendong Zhao
|
Yunfei Yang
|
Xin Zhao
|
Jingcheng He
Findings of the Association for Computational Linguistics: EMNLP 2025
Texts generated by large language models (LLMs) are increasingly widespread online. Due to the lack of effective attribution mechanisms, the enforcement of copyright and the prevention of misuse remain significant challenges in the context of LLM-generated content. LLMs watermark emerges as a crucial technology to trace the source of AI-generated content. However, most existing watermarking methods reduce the fidelity of semantics. To address this issue, this paper introduces a novel watermarking framework. To enhance the fidelity of semantics, we propose low-entropy POS-guided token partitioning mechanism and z-score-driven dynamic bias mechanism. Moreover, to enhance the robustness against potential bias sparsity exploitation attack, we propose a relative position encoding (RPE) mechanism, which can uniformly distribute bias in the generated text. Evaluated across 6 baselines, 4 tasks, and 5 LLMs under 8 attacks, compared to the KGW, our watermark improves semantic fidelity by 24.53% (RC-PPL) and robustness by 3.75% (F1). Our code is publicly available, facilitating reproducibility in LLM watermarking research.
2015
A Computationally Efficient Algorithm for Learning Topical Collocation Models
Zhendong Zhao
|
Lan Du
|
Benjamin Börschinger
|
John K Pate
|
Massimiliano Ciaramita
|
Mark Steedman
|
Mark Johnson
Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
Search
Fix author
Co-authors
- Benjamin Börschinger 1
- Xiaojun Chen 1
- Massimiliano Ciaramita 1
- Lan Du 1
- Jingcheng He 1
- show all...