An Assessment of Multistage Reward Function Design for Deep Reinforcement Learning-Based Microgrid Energy Management

dc.contributor.authorGoh, Hui Hwangen
dc.contributor.authorHuang, Yifengen
dc.contributor.authorLim, Chee Shenen
dc.contributor.authorZhang, Dongdongen
dc.contributor.authorLiu, Huien
dc.contributor.authorDai, Weien
dc.contributor.authorKurniawan, Tonni Agustionoen
dc.contributor.authorRahman, Saifuren
dc.date.accessioned2024-03-27T19:10:38Zen
dc.date.available2024-03-27T19:10:38Zen
dc.date.issued2022-06-01en
dc.description.abstractReinforcement learning based energy management strategy has been an active research subject in the past few years. Different from the baseline reward function (BRF), the work proposes and investigates a multi-stage reward mechanism (MSRM) that scores the agent's step and final performance during training and returns it to the agent in real time as a reward. MSRM will also improve the agent's training through expert intervention which aims to prevent the agent from being trapped in sub-optimal strategies. The energy management performance considered by MSRM-based algorithm includes the energy balance, economic cost, and reliability. The reward function is assessed in conjunction with two deep reinforcement learning algorithms: double deep Q-learning network (DDQN) and policy gradient (PG). Upon benchmarking with BRF, the numerical simulation shows that MSRM tends to improve the convergence characteristic, reduce the explained variance, and reduce the tendency of the agent being trapped in suboptimal strategies. In addition, the methods have been assessed with MPC-based energy management strategies in terms of relative cost, self-balancing rate, and computational time. The assessment concludes that, in the given context, PG-MSRM has the best overall performance.en
dc.description.versionAccepted versionen
dc.format.extentPages 4300-4311en
dc.format.extent12 page(s)en
dc.format.mimetypeapplication/pdfen
dc.identifier.doihttps://doi.org/10.1109/TSG.2022.3179567en
dc.identifier.eissn1949-3061en
dc.identifier.issn1949-3053en
dc.identifier.issue6en
dc.identifier.orcidRahman, Saifur [0000-0001-6226-8406]en
dc.identifier.urihttps://hdl.handle.net/10919/118472en
dc.identifier.volume13en
dc.language.isoenen
dc.publisherIEEEen
dc.rightsIn Copyrighten
dc.rights.urihttp://rightsstatements.org/vocab/InC/1.0/en
dc.subjectMicrogridsen
dc.subjectEnergy managementen
dc.subjectReal-time systemsen
dc.subjectCostsen
dc.subjectPrediction algorithmsen
dc.subjectTrainingen
dc.subjectConvergenceen
dc.subjectMicrogrid energy managementen
dc.subjectdeep reinforcement learningen
dc.subjectreward functionen
dc.subjectoptimal schedulingen
dc.titleAn Assessment of Multistage Reward Function Design for Deep Reinforcement Learning-Based Microgrid Energy Managementen
dc.title.serialIEEE Transactions on Smart Griden
dc.typeArticle - Refereeden
dc.type.dcmitypeTexten
dc.type.otherArticleen
dc.type.otherJournalen
pubs.organisational-group/Virginia Techen
pubs.organisational-group/Virginia Tech/Engineeringen
pubs.organisational-group/Virginia Tech/Engineering/Advanced Research Instituteen
pubs.organisational-group/Virginia Tech/Engineering/Electrical and Computer Engineeringen
pubs.organisational-group/Virginia Tech/All T&R Facultyen
pubs.organisational-group/Virginia Tech/Engineering/COE T&R Facultyen

Files

Original bundle
Now showing 1 - 1 of 1
Name:
Goh-TSG-An Assessment of Multi-Stage.docx
Size:
3.31 MB
Format:
Microsoft Word XML
Description:
Accepted version
License bundle
Now showing 1 - 1 of 1
Name:
license.txt
Size:
1.5 KB
Format:
Plain Text
Description: