Cooperative price-based demand response program for multiple aggregators based on multi-agent reinforcement learning and Shapley-value

Téléchargements

Téléchargements par mois depuis la dernière année

Fraija, A., Henao, N., Agbossou, K., Kelouwani, S. et Fournier, M. (2024). Cooperative price-based demand response program for multiple aggregators based on multi-agent reinforcement learning and Shapley-value. Sustainable Energy, Grids and Networks, 40 . Article 101560. ISSN 2352-4677 DOI 10.1016/j.segan.2024.101560

[thumbnail of AGBOSSOU_K_40_ED.pdf]
Prévisualisation
PDF
Télécharger (2MB) | Prévisualisation

Résumé

Abstract

Demand response (DR) plays an essential role in power system management. To facilitate the implementation of these techniques, many aggregators have appeared in response as new mediating entities in the electricity market. These actors exploit the technologies to engage customers in DR programs, offering grid services like load scheduling. However, the growing number of aggregators has become a new challenge, making it difficult for utilities to manage the load scheduling problem. This paper presents a multi-agent reinforcement Learning (MARL) approach to a price-based DR program for multiple aggregators. A dynamic pricing scheme based on discounts is proposed to encourage residential customers to change their consumption patterns. This strategy is based on a cooperative framework for a set of DR Aggregators (DRAs). The DRAs take advantage of a reward offered by a Distribution System Operator (DSO) for performing a peak-shaving over the total system aggregated demand. Furthermore, a Shapley-Value-based reward sharing mechanism is implemented to fairly determine the individual contribution and calculate the individual reward for each DRA. Simulation results verify the merits of the proposed model for a multi-aggregator system, improving DRAs’ pricing strategies considering the overall objectives of the system. Consumption peaks were managed by reducing the Peak-to-Average Ratio (PAR) by 15%, and the MARL mechanism's performance was improved in terms of reward function maximization and convergence time, the latter being reduced by 29%.

Type de document: Article
Mots-clés libres: Demand response Demand response aggregator Dynamic pricing Multi-agent reinforcement learning Shapley-value
Date de dépôt: 22 janv. 2026 15:58
Dernière modification: 22 janv. 2026 15:58
Version du document déposé: Version officielle de l'éditeur
URI: https://depot-e.uqtr.ca/id/eprint/12551

Actions (administrateurs uniquement)

Éditer la notice Éditer la notice