Personalized Author Obfuscation with Large Language Models

Abstract
In this paper, we investigate the efficacy of large language models (LLMs) in obfuscating authorship by paraphrasing and altering writing styles. Rather than adopting a holistic approach that evaluates performance across the entire dataset, we focus on user-wise performance to analyze how obfuscation effectiveness varies across individual authors. While LLMs are generally effective, we observe a bimodal distribution of efficacy, with performance varying significantly across users. To address this, we propose a personalized prompting method that outperforms standard prompting techniques and partially mitigates the bimodality issue.
View on arXiv@article{shokri2025_2505.12090, title={ Personalized Author Obfuscation with Large Language Models }, author={ Mohammad Shokri and Sarah Ita Levitan and Rivka Levitan }, journal={arXiv preprint arXiv:2505.12090}, year={ 2025 } }
Comments on this paper