Multi-Objective Linguistic Control of Large Language Models
Dang Nguyen, Jiuhai Chen, Tianyi Zhou
TL;DR
This work tackles the challenge of controlling multiple linguistic complexities in large language models. It introduces Multi-Control Tuning (MCTune), a simple, data-efficient approach that appends a vector of linguistic complexity features to instruction inputs and trains models to generate responses conditioned on selected controls. By finetuning LLaMA2-7B on existing instruction-tuning data (Alpaca-GPT4 and WizardLM) with randomized subsets of controls, MCTune achieves substantial improvements in controllability across multiple features while maintaining or even improving generation quality, as evidenced by MT-Bench and pairwise evaluations. The method leverages off-the-shelf data, a Gaussian-based sampling strategy for evaluation, and a sigma hyperparameter to adjust difficulty, demonstrating practical, scalable multi-objective linguistic control with broad implications for personalized and adaptive AI systems.
Abstract
Large language models (LLMs), despite their breakthroughs on many challenging benchmark tasks, lean to generate verbose responses and lack the controllability of output complexity, which is usually preferred by human users in practice. In this paper, we study how to precisely control multiple linguistic complexities of LLM output by finetuning using off-the-shelf data. To this end, we propose multi-control tuning (MCTune), which includes multiple linguistic complexity values of ground-truth responses as controls in the input for instruction tuning. We finetune LLaMA2-7B on Alpaca-GPT4 and WizardLM datasets. Evaluations on widely used benchmarks demonstrate that our method does not only improve LLMs' multi-complexity controllability substantially but also retains or even enhances the quality of the responses as a side benefit.
