Learning in Stackelberg Games with Application ...
Type de document :
Autre communication scientifique (congrès sans actes - poster - séminaire...): Communication dans un congrès avec actes
Titre :
Learning in Stackelberg Games with Application to Strategic Bidding in the Electricity Market
Auteur(s) :
Morri, Francesco [Auteur correspondant]
Integrated Optimization with Complex Structure [INOCS]
Inria Lille - Nord Europe
Le Cadre, Hélène [Auteur]
Integrated Optimization with Complex Structure [INOCS]
Inria Lille - Nord Europe
Brotcorne, Luce [Auteur]
Integrated Optimization with Complex Structure [INOCS]
Inria Lille - Nord Europe
Gruet, Pierre [Auteur]
EDF R&D [EDF R&D]
Integrated Optimization with Complex Structure [INOCS]
Inria Lille - Nord Europe
Le Cadre, Hélène [Auteur]
Integrated Optimization with Complex Structure [INOCS]
Inria Lille - Nord Europe
Brotcorne, Luce [Auteur]
Integrated Optimization with Complex Structure [INOCS]
Inria Lille - Nord Europe
Gruet, Pierre [Auteur]
EDF R&D [EDF R&D]
Titre de la manifestation scientifique :
EEM24
Ville :
Istanbul
Pays :
Turquie
Date de début de la manifestation scientifique :
2024-06-10
Mot(s)-clé(s) en anglais :
Energy Market
Multi Agent Reinforcement Learning
Market Simulation
Multi Agent Reinforcement Learning
Market Simulation
Discipline(s) HAL :
Informatique [cs]/Apprentissage [cs.LG]
Résumé en anglais : [en]
We formulate a two-stage electricity market involving conventional and renewable producers strategically bidding in the day-ahead market, to maximize their profits while anticipating the market clearing performed by an ...
Lire la suite >We formulate a two-stage electricity market involving conventional and renewable producers strategically bidding in the day-ahead market, to maximize their profits while anticipating the market clearing performed by an Independent System Operator (ISO), as a multi-leader single follower Stackelberg game. In this game, producers are interpreted as leaders, while the ISO acts as a follower.To compute an equilibrium, the classical approach is to cast the Stackelberg game as a Generalized Nash Game (GNG), replacing the ISO's optimization problem by its KKT constraints. To solve this reformulated problem, we can either rely on the Gauss-Seidel Best-Response method (GS-BR), or, on the Alternating Direction Method of Multipliers (ADMM). However, both approaches are implemented in a centralized setting since they require the existence of a coordinator which keeps track of the history of agents' strategies and sequential updates, or, is responsible for the Lagrange multiplier updates following the augmented Lagrangian.To allow the agents to selfishly optimize their utility functions in a decentralized setting, we introduce a variant of an actor-critic Multi-Agent deep Reinforcement Learning (MARL) algorithm with provable convergence.Our algorithm is innovative in that it allows different levels of coordination among the actors and the critic, thus capturing different information structures of the Stackelberg game. We conclude this work by comparing GS-BR and ADMM, both used as benchmark, to the MARL, on a dataset from the French electricity market, relying on metrics such as the efficiency loss and the accuracy of the solution.Lire moins >
Lire la suite >We formulate a two-stage electricity market involving conventional and renewable producers strategically bidding in the day-ahead market, to maximize their profits while anticipating the market clearing performed by an Independent System Operator (ISO), as a multi-leader single follower Stackelberg game. In this game, producers are interpreted as leaders, while the ISO acts as a follower.To compute an equilibrium, the classical approach is to cast the Stackelberg game as a Generalized Nash Game (GNG), replacing the ISO's optimization problem by its KKT constraints. To solve this reformulated problem, we can either rely on the Gauss-Seidel Best-Response method (GS-BR), or, on the Alternating Direction Method of Multipliers (ADMM). However, both approaches are implemented in a centralized setting since they require the existence of a coordinator which keeps track of the history of agents' strategies and sequential updates, or, is responsible for the Lagrange multiplier updates following the augmented Lagrangian.To allow the agents to selfishly optimize their utility functions in a decentralized setting, we introduce a variant of an actor-critic Multi-Agent deep Reinforcement Learning (MARL) algorithm with provable convergence.Our algorithm is innovative in that it allows different levels of coordination among the actors and the critic, thus capturing different information structures of the Stackelberg game. We conclude this work by comparing GS-BR and ADMM, both used as benchmark, to the MARL, on a dataset from the French electricity market, relying on metrics such as the efficiency loss and the accuracy of the solution.Lire moins >
Langue :
Anglais
Comité de lecture :
Oui
Audience :
Internationale
Vulgarisation :
Non
Collections :
Source :
Fichiers
- document
- Accès libre
- Accéder au document
- EEM24_Paper_with_names.pdf
- Accès libre
- Accéder au document