Calculate Bellman values throughout iterations of Antares simulation and DP Each simulation leads to a new reward estimation, which leads to new water values, which leads to the off-line calculation in R of an optimal trajectory, which leads to new controls to be evaluated which leads to a new simulation
Calculate Bellman values throughout iterations of Antares simulation and DP Each simulation leads to a new reward estimation, which leads to new water values, which leads to the off-line calculation in R of an optimal trajectory, which leads to new controls to be evaluated which leads to a new simulation
Calculate an optimal trajectory for the reservoir levels based on water values taking into account the mean inflow, used in calculateBellmanWithIterativeSimulations
Get objective values of the optimization problem of each week and each scenario for a given simulation simu, mainly used in get_Reward() to build reward functions.