ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2510.02342
60
0

CATMark: A Context-Aware Thresholding Framework for Robust Cross-Task Watermarking in Large Language Models

27 September 2025
Yu Zhang
Shuliang Liu
Xu Yang
Xuming Hu
    WaLM
ArXiv (abs)PDFHTML
Main:7 Pages
11 Figures
Bibliography:3 Pages
3 Tables
Appendix:16 Pages
Abstract

Watermarking algorithms for Large Language Models (LLMs) effectively identify machine-generated content by embedding and detecting hidden statistical features in text. However, such embedding leads to a decline in text quality, especially in low-entropy scenarios where performance needs improvement. Existing methods that rely on entropy thresholds often require significant computational resources for tuning and demonstrate poor adaptability to unknown or cross-task generation scenarios. We propose \textbf{C}ontext-\textbf{A}ware \textbf{T}hreshold watermarking (\myalgo\myalgo\myalgo), a novel framework that dynamically adjusts watermarking intensity based on real-time semantic context. \myalgo\myalgo\myalgo partitions text generation into semantic states using logits clustering, establishing context-aware entropy thresholds that preserve fidelity in structured content while embedding robust watermarks. Crucially, it requires no pre-defined thresholds or task-specific tuning. Experiments show \myalgo\myalgo\myalgo improves text quality in cross-tasks without sacrificing detection accuracy.

View on arXiv
Comments on this paper