CIE: Controlling Language Model Text Generations Using Continuous Signals

Aligning language models with user intent is becoming increasingly relevant to enhance user experience. This calls for designing methods that can allow users to control the properties of the language that LMs generate. For example, controlling the length of the generation, the complexity of the language that gets chosen, the sentiment, tone, etc. Most existing work attempts to integrate users' control by conditioning LM generations on natural language prompts or discrete control signals, which are often brittle and hard to scale. In this work, we are interested in \textit{continuous} control signals, ones that exist along a spectrum that can't easily be captured in a natural language prompt or via existing techniques in conditional generation. Through a case study in controlling the precise response-length of generations produced by LMs, we demonstrate how after fine-tuning, behaviors of language models can be controlled via continuous signals -- as vectors that are interpolated between a "low" and a "high" token embedding. Our method more reliably exerts response-length control than in-context learning methods or fine-tuning methods that represent the control signal as a discrete signal. Our full open-sourced code and datasets are available atthis https URL.
View on arXiv@article{samuel2025_2505.13448, title={ CIE: Controlling Language Model Text Generations Using Continuous Signals }, author={ Vinay Samuel and Harshita Diddee and Yiming Zhang and Daphne Ippolito }, journal={arXiv preprint arXiv:2505.13448}, year={ 2025 } }