From Static to Adaptive: Dynamic Cost Function Weight Adaptation in Hierarchical Reinforcement Learning for Sustainable 6G Radio Access Networks

dc.contributor.authorViana Fonseca Abreu, Jeffersonen
dc.contributor.committeechairKibilda, Jaceken
dc.contributor.committeememberMidkiff, Scott F.en
dc.contributor.committeememberPereira da Silva, Luiz Antonioen
dc.contributor.departmentElectrical and Computer Engineeringen
dc.date.accessioned2025-12-20T09:00:22Zen
dc.date.available2025-12-20T09:00:22Zen
dc.date.issued2025-12-19en
dc.description.abstractThe rapid growth of mobile network traffic and the densification required for 6G networks significantly increase energy consumption, with base stations (BS) accounting for up to 70% of total network energy use. Energy-efficient BS switching has therefore become a critical research focus. Traditional solutions rely on static thresholds or fixed cost function weights, limiting adaptability in dynamic environments. This thesis investigates how cost function design and weight adaptation influence the trade-off between energy consumption and Quality of Service (QoS) degradation in Deep Reinforcement Learning (DRL)-based BS switching. Using a realistic spatio-temporal dataset, we show that static cost weights lead to suboptimal performance under varying traffic conditions. To address this, we propose a Hierarchical Reinforcement Learning (HRL) architecture in which a high-level controller dynamically selects low-level policies trained with different cost function weights. Experimental results demonstrate that the proposed HRL approach achieves up to 64% energy reduction—improving by 5% over the static DRL baseline—while maintaining acceptable QoS levels. These findings highlight the potential of hierarchical control and adaptive weighting in achieving scalable, sustainable 6G Radio Access Networks operations.en
dc.description.abstractgeneralWireless communications systems (used by devices such as cellphones, tablets, and cars, among others) utilize specialized equipment to send signals and information over the air. This set of equipment, known as Base Stations, serves as a bridge between devices and the internet. To serve the devices, the Base Stations run continuously, using a large part of the energy consumed in a wireless communications system. As the demand for wireless connectivity increases, it is expected that the number of Base Stations will continue to grow, further increasing energy consumption. One straightforward way to save energy is to temporarily turn off some base stations during periods when fewer devices are using the network. However, deciding when to turn them on or off is difficult because the number of active devices, or traffic, changes throughout the day. This thesis explores whether computer programs that are trained and learn from experience can make better decisions about when to power base stations on or off. These programs learn by trying actions and receiving feedback on whether the choice was good or bad. Usually, this feedback is designed before the computer program's learning process and remains fixed throughout all the computer program's interactions. Our work demonstrates that a computer program trained using fixed feedback is ineffective when user demand fluctuates. We propose a two-level decision system in which one program selects, based on the user demand, which other computer programs will be active. Each of the eligible computer programs was trained with different feedback, therefore having different priorities regarding saving energy or maintaining service quality. This approach reduces energy use by up to 64% while keeping service performance at an acceptable level. These results suggest that learning-based decision systems can help future wireless networks operate more efficiently and sustainably.en
dc.description.degreeMaster of Scienceen
dc.format.mediumETDen
dc.identifier.othervt_gsexam:45243en
dc.identifier.urihttps://hdl.handle.net/10919/140536en
dc.language.isoenen
dc.publisherVirginia Techen
dc.rightsIn Copyrighten
dc.rights.urihttp://rightsstatements.org/vocab/InC/1.0/en
dc.subjectEnergy efficiencyen
dc.subjectBase Station Switchingen
dc.subjectHierarchical Reinforcement Learningen
dc.titleFrom Static to Adaptive: Dynamic Cost Function Weight Adaptation in Hierarchical Reinforcement Learning for Sustainable 6G Radio Access Networksen
dc.typeThesisen
thesis.degree.disciplineComputer Engineeringen
thesis.degree.grantorVirginia Polytechnic Institute and State Universityen
thesis.degree.levelmastersen
thesis.degree.nameMaster of Scienceen

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Viana_Fonseca_Abreu_J_T_2025.pdf
Size:
511.28 KB
Format:
Adobe Portable Document Format

Collections