/
L2RPNReward.py
46 lines (37 loc) · 1.86 KB
/
L2RPNReward.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
# Copyright (c) 2019-2020, RTE (https://www.rte-france.com)
# See AUTHORS.txt
# This Source Code Form is subject to the terms of the Mozilla Public License, version 2.0.
# If a copy of the Mozilla Public License, version 2.0 was not distributed with this file,
# you can obtain one at http://mozilla.org/MPL/2.0/.
# SPDX-License-Identifier: MPL-2.0
# This file is part of Grid2Op, Grid2Op a testbed platform to model sequential decision making in power systems.
import numpy as np
from grid2op.Reward.BaseReward import BaseReward
from grid2op.dtypes import dt_float
class L2RPNReward(BaseReward):
"""
This is the historical :class:`BaseReward` used for the Learning To Run a Power Network competition.
See `L2RPN <https://l2rpn.chalearn.org/>`_ for more information.
"""
def __init__(self):
BaseReward.__init__(self)
def initialize(self, env):
self.reward_min = dt_float(0.0)
self.reward_max = dt_float(env.backend.n_line)
def __call__(self, action, env, has_error, is_done, is_illegal, is_ambiguous):
if not is_done and not has_error:
line_cap = self.__get_lines_capacity_usage(env)
res = np.sum(line_cap)
else:
# no more data to consider, no powerflow has been run, reward is what it is
res = self.reward_min
return res
@staticmethod
def __get_lines_capacity_usage(env):
ampere_flows = np.abs(env.backend.get_line_flow(), dtype=dt_float)
thermal_limits = np.abs(env.get_thermal_limit(), dtype=dt_float)
thermal_limits += 1e-1 # for numerical stability
relative_flow = np.divide(ampere_flows, thermal_limits, dtype=dt_float)
x = np.minimum(relative_flow, dt_float(1.0))
lines_capacity_usage_score = np.maximum(dt_float(1.0) - x ** 2, np.zeros(x.shape, dtype=dt_float))
return lines_capacity_usage_score