Inducing Personality in LLM-Based Honeypot Agents: Measuring the Effect on Human-Like Agenda Generation

Abstract
This paper presents SANDMAN, an architecture for cyber deception that leverages Language Agents to emulate convincing human simulacra. Our 'Deceptive Agents' serve as advanced cyber decoys, designed for high-fidelity engagement with attackers by extending the observation period of attack behaviours. Through experimentation, measurement, and analysis, we demonstrate how a prompt schema based on the five-factor model of personality systematically induces distinct 'personalities' in Large Language Models. Our results highlight the feasibility of persona-driven Language Agents for generating diverse, realistic behaviours, ultimately improving cyber deception strategies.
View on arXiv@article{newsham2025_2503.19752, title={ Inducing Personality in LLM-Based Honeypot Agents: Measuring the Effect on Human-Like Agenda Generation }, author={ Lewis Newsham and Ryan Hyland and Daniel Prince }, journal={arXiv preprint arXiv:2503.19752}, year={ 2025 } }
Comments on this paper