64
0

V-VAE: A Variational Auto Encoding Framework Towards Fine-Grained Control over Human-Like Chat

Main:7 Pages
7 Figures
Bibliography:3 Pages
5 Tables
Appendix:4 Pages
Abstract

With the continued proliferation of Large Language Model (LLM) based chatbots, there is a growing demand for generating responses that are not only linguistically fluent but also consistently aligned with persona-specific traits in conversations. However, existing role-play and persona-based chat approaches rely heavily on static role descriptions, coarse-grained signal space, and low-quality synthetic data, which fail to capture dynamic fine-grained details in human-like chat. Human-like chat requires modeling subtle latent traits, such as emotional tone, situational awareness, and evolving personality, which are difficult to predefine and cannot be easily learned from synthetic or distillation-based data. To address these limitations, we propose a Verbal Variational Auto-Encoding (V-VAE) framework, containing a variational auto-encoding module and fine-grained control space which dynamically adapts dialogue behaviour based on fine-grained, interpretable latent variables across talking style, interaction patterns, and personal attributes. We also construct a high-quality dataset, HumanChatData, and benchmark HumanChatBench to address the scarcity of high-quality data in the human-like domain. Experiments show that LLMs based on V-VAE consistently outperform standard baselines on HumanChatBench and DialogBench, which further demonstrates the effectiveness of V-VAE and HumanChatData.

View on arXiv
@article{lin2025_2506.01524,
  title={ V-VAE: A Variational Auto Encoding Framework Towards Fine-Grained Control over Human-Like Chat },
  author={ Qi Lin and Weikai Xu and Lisi Chen and Bin Dai },
  journal={arXiv preprint arXiv:2506.01524},
  year={ 2025 }
}
Comments on this paper