Temporal Prototype-Aware Learning for Active Voltage Control on Power Distribution Networks
Feiyang Xu, Shunyu Liu, Yunpeng Qing, Yihe Zhou, Yuwen Wang, Mingli Song
TL;DR
This paper tackles the challenge of sustaining effective active voltage control on power distribution networks under long-term temporal distribution shifts, using short-term training trajectories. It proposes Temporal Prototype-Aware learning (TPA), which combines a multi-scale dynamic encoder (a stacked transformer) to capture multi-timescale temporal dependencies with a temporal prototype-aware policy that matches ongoing states to seasonal prototypes for adaptive control. Key contributions include the design of 24 seasonal prototypes, a prototype-based retrieval mechanism, and ablation analyses showing the importance of short-term memory and temporal priors, along with transferability across PDN sizes. Empirical results on MAPDN benchmarks (141- and 322-bus) demonstrate that TPA outperforms state-of-the-art MARL baselines in both singular diurnal and longer operational cycles, offering improved controllable rate and reduced reactive-power loss, with practical implications for scalable, time-adaptive AVC in real-world grids.
Abstract
Active Voltage Control (AVC) on the Power Distribution Networks (PDNs) aims to stabilize the voltage levels to ensure efficient and reliable operation of power systems. With the increasing integration of distributed energy resources, recent efforts have explored employing multi-agent reinforcement learning (MARL) techniques to realize effective AVC. Existing methods mainly focus on the acquisition of short-term AVC strategies, i.e., only learning AVC within the short-term training trajectories of a singular diurnal cycle. However, due to the dynamic nature of load demands and renewable energy, the operation states of real-world PDNs may exhibit significant distribution shifts across varying timescales (e.g., daily and seasonal changes). This can render those short-term strategies suboptimal or even obsolete when performing continuous AVC over extended periods. In this paper, we propose a novel temporal prototype-aware learning method, abbreviated as TPA, to learn time-adaptive AVC under short-term training trajectories. At the heart of TPA are two complementary components, namely multi-scale dynamic encoder and temporal prototype-aware policy, that can be readily incorporated into various MARL methods. The former component integrates a stacked transformer network to learn underlying temporal dependencies at different timescales of the PDNs, while the latter implements a learnable prototype matching mechanism to construct a dedicated AVC policy that can dynamically adapt to the evolving operation states. Experimental results on the AVC benchmark with different PDN sizes demonstrate that the proposed TPA surpasses the state-of-the-art counterparts not only in terms of control performance but also by offering model transferability. Our code is available at https://github.com/Canyizl/TPA-for-AVC.
