MT2ST: Adaptive Multi-Task to Single-Task Learning
Dong Liu, Yanxuan Yu
TL;DR
MT2ST addresses the trade-off between generalization in multi-task learning and task-specific precision in single-task learning by introducing an adaptive training framework that transitions from multi-task to single-task optimization. It combines a shared encoder with two transition mechanisms—Diminish (soft, exponential decay of auxiliary losses) and Switch (hard, epoch-based transition to the main task)—to realize early generalization followed by late specialization. The framework is demonstrated across representation learning, transformers, and diffusion models, showing substantial training-time reductions and competitive or improved accuracy, including up to 67% speed-ups in embedding tasks and notable gains in vision and multimodal settings. MT2ST offers a practical, architecture-agnostic pathway for efficient, scalable task-oriented representation learning with potential extensions to vision and multimodal domains.
Abstract
Efficient machine learning (ML) has become increasingly important as models grow larger and data volumes expand. In this work, we address the trade-off between generalization in multi-task learning (MTL) and precision in single-task learning (STL) by introducing the Multi-Task to Single-Task (MT2ST) framework. MT2ST is designed to enhance training efficiency and accuracy in multi-modal tasks, showcasing its value as a practical application of efficient ML.
