Automating Security Audit Using Large Language Model based Agent: An Exploration Experiment

In the current rapidly changing digital environment, businesses are under constant stress to ensure that their systems are secured. Security audits help to maintain a strong security posture by ensuring that policies are in place, controls are implemented, gaps are identified for cybersecurity risks mitigation. However, audits are usually manual, requiring much time and costs. This paper looks at the possibility of developing a framework to leverage Large Language Models (LLMs) as an autonomous agent to execute part of the security audit, namely with the field audit. password policy compliance for Windows operating system. Through the conduct of an exploration experiment of using GPT-4 with Langchain, the agent executed the audit tasks by accurately flagging password policy violations and appeared to be more efficient than traditional manual audits. Despite its potential limitations in operational consistency in complex and dynamic environment, the framework suggests possibilities to extend further to real-time threat monitoring and compliance checks.
View on arXiv@article{chin2025_2505.10732, title={ Automating Security Audit Using Large Language Model based Agent: An Exploration Experiment }, author={ Jia Hui Chin and Pu Zhang and Yu Xin Cheong and Jonathan Pan }, journal={arXiv preprint arXiv:2505.10732}, year={ 2025 } }