TY - GEN
T1 - Reinforcement learning for cyber-physical security assessment of power systems
AU - Liu, Xiaorui
AU - Konstantinou, Charalambos
N1 - Generated from Scopus record by KAUST IRTS on 2022-09-13
PY - 2019/6/1
Y1 - 2019/6/1
N2 - The protection of power systems is of paramount significance for the supply of electricity. Contingency analysis allows to access the impact of power grid components failures. Typically, power systems are designed to handle N-2 contingencies. Existing algorithms mainly focus on performance and computational efficiency. There has been little effort in designing contingency methods from a cybersecurity perspective. To address this limitation, we study contingency analysis in the context of power system planning and operation towards cyber-physical security assessment. The proposed methodology considers attackers transitions in the network based on the N-2 critical contingency pairs. We develop an online reinforcement Q -learning scheme to solve a Markov decision process that models adversarial actions. In this model, the adversary aims to maximize the cumulative reward before making any action and learns adaptively how to optimize the attack strategies. We validate and test the algorithm on eleven literature-based and synthetic power grid test cases.
AB - The protection of power systems is of paramount significance for the supply of electricity. Contingency analysis allows to access the impact of power grid components failures. Typically, power systems are designed to handle N-2 contingencies. Existing algorithms mainly focus on performance and computational efficiency. There has been little effort in designing contingency methods from a cybersecurity perspective. To address this limitation, we study contingency analysis in the context of power system planning and operation towards cyber-physical security assessment. The proposed methodology considers attackers transitions in the network based on the N-2 critical contingency pairs. We develop an online reinforcement Q -learning scheme to solve a Markov decision process that models adversarial actions. In this model, the adversary aims to maximize the cumulative reward before making any action and learns adaptively how to optimize the attack strategies. We validate and test the algorithm on eleven literature-based and synthetic power grid test cases.
UR - https://ieeexplore.ieee.org/document/8810568/
UR - http://www.scopus.com/inward/record.url?scp=85072349092&partnerID=8YFLogxK
U2 - 10.1109/PTC.2019.8810568
DO - 10.1109/PTC.2019.8810568
M3 - Conference contribution
SN - 9781538647226
BT - 2019 IEEE Milan PowerTech, PowerTech 2019
PB - Institute of Electrical and Electronics Engineers Inc.
ER -