当前位置: X-MOL 学术Journal of Modern Power Systems and Clean Energy › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Power System Flow Adjustment and Sample Generation Based on Deep Reinforcement Learning
Journal of Modern Power Systems and Clean Energy ( IF 6.3 ) Pub Date : 2020-12-02 , DOI: 10.35833/mpce.2020.000240
Shuang Wu , Wei Hu , Zongxiang Lu , Yujia Gu , Bei Tian , Hongqiang Li

With the increasing complexity of power system structures and the increasing penetration of renewable energy, the number of possible power system operation modes increases dramatically. It is difficult to make manual power flow adjustments to establish an initial convergent power flow that is suitable for operation mode analysis. At present, problems of low efficiency and long time consumption are encountered in the formulation of operation modes, resulting in a very limited number of generated operation modes. In this paper, we propose an intelligent power flow adjustment and generation model based on a deep network and reinforcement learning. First, a discriminator is trained to judge the power flow convergence, and the output of this discriminator is used to construct a value function. Then, the reinforcement learning method is adopted to learn a strategy for power flow convergence adjustment. Finally, a large number of convergent power flow samples are generated using the learned adjustment strategy. Compared with the traditional flow adjustment method, the proposed method has significant advantages that the learning of the power flow adjustment strategy does not depend on the parameters of the power system model. Therefore, this strategy can be automatically learned without manual intervention, which allows a large number of different operation modes to be efficiently formulated. The verification results of a case study show that the proposed method can independently learn a power flow adjustment strategy and generate various convergent power flows.

中文翻译:

基于深度强化学习的电力系统流量调整与样本生成

随着电力系统结构的复杂性增加和可再生能源的普及,可能的电力系统运行模式数量急剧增加。难以进行手动功率潮流调整​​以建立适合于操作模式分析的初始会聚功率潮流。当前,在操作模式的制定中遇到了效率低和时间消耗长的问题,导致生成的操作模式的数量非常有限。本文提出了一种基于深度网络和强化学习的智能潮流调整与生成模型。首先,训练一个鉴别器来判断潮流收敛,然后使用该鉴别器的输出来构造一个值函数。然后,采用强化学习法学习潮流收敛调整策略。最后,使用学习到的调整策略生成大量收敛的潮流样本。与传统的潮流调整方法相比,该方法具有明显的优势,即潮流调整策略的学习不依赖于电力系统模型的参数。因此,无需手动干预即可自动学习该策略,从而可以有效地制定大量不同的操作模式。案例研究的验证结果表明,该方法可以独立学习潮流调整策略,并产生各种收敛的潮流。使用学习到的调整策略会生成大量收敛的潮流样本。与传统的潮流调整方法相比,该方法具有明显的优势,即潮流调整策略的学习不依赖于电力系统模型的参数。因此,无需手动干预即可自动学习该策略,从而可以有效地制定大量不同的操作模式。案例研究的验证结果表明,该方法可以独立学习潮流调整策略,并产生各种收敛的潮流。使用学习到的调整策略可生成大量收敛的潮流样本。与传统的潮流调整方法相比,该方法具有明显的优势,即潮流调整策略的学习不依赖于电力系统模型的参数。因此,无需手动干预即可自动学习此策略,从而可以有效地制定大量不同的操作模式。案例研究的验证结果表明,该方法可以独立学习潮流调整策略,并产生各种收敛的潮流。该方法具有明显的优势,即潮流调整策略的学习不依赖于电力系统模型的参数。因此,无需手动干预即可自动学习该策略,从而可以有效地制定大量不同的操作模式。案例研究的验证结果表明,该方法可以独立学习潮流调整策略,并产生各种收敛的潮流。该方法具有明显的优势,即潮流调整策略的学习不依赖于电力系统模型的参数。因此,无需手动干预即可自动学习此策略,从而可以有效地制定大量不同的操作模式。案例研究的验证结果表明,该方法可以独立学习潮流调整策略,并产生各种收敛的潮流。
更新日期:2020-12-04
down
wechat
bug