From Static to Adaptive: Dynamic Cost Function Weight Adaptation in Hierarchical Reinforcement Learning for Sustainable 6G Radio Access Networks
| dc.contributor.author | Viana Fonseca Abreu, Jefferson | en |
| dc.contributor.committeechair | Kibilda, Jacek | en |
| dc.contributor.committeemember | Midkiff, Scott F. | en |
| dc.contributor.committeemember | Pereira da Silva, Luiz Antonio | en |
| dc.contributor.department | Electrical and Computer Engineering | en |
| dc.date.accessioned | 2025-12-20T09:00:22Z | en |
| dc.date.available | 2025-12-20T09:00:22Z | en |
| dc.date.issued | 2025-12-19 | en |
| dc.description.abstract | The rapid growth of mobile network traffic and the densification required for 6G networks significantly increase energy consumption, with base stations (BS) accounting for up to 70% of total network energy use. Energy-efficient BS switching has therefore become a critical research focus. Traditional solutions rely on static thresholds or fixed cost function weights, limiting adaptability in dynamic environments. This thesis investigates how cost function design and weight adaptation influence the trade-off between energy consumption and Quality of Service (QoS) degradation in Deep Reinforcement Learning (DRL)-based BS switching. Using a realistic spatio-temporal dataset, we show that static cost weights lead to suboptimal performance under varying traffic conditions. To address this, we propose a Hierarchical Reinforcement Learning (HRL) architecture in which a high-level controller dynamically selects low-level policies trained with different cost function weights. Experimental results demonstrate that the proposed HRL approach achieves up to 64% energy reduction—improving by 5% over the static DRL baseline—while maintaining acceptable QoS levels. These findings highlight the potential of hierarchical control and adaptive weighting in achieving scalable, sustainable 6G Radio Access Networks operations. | en |
| dc.description.abstractgeneral | Wireless communications systems (used by devices such as cellphones, tablets, and cars, among others) utilize specialized equipment to send signals and information over the air. This set of equipment, known as Base Stations, serves as a bridge between devices and the internet. To serve the devices, the Base Stations run continuously, using a large part of the energy consumed in a wireless communications system. As the demand for wireless connectivity increases, it is expected that the number of Base Stations will continue to grow, further increasing energy consumption. One straightforward way to save energy is to temporarily turn off some base stations during periods when fewer devices are using the network. However, deciding when to turn them on or off is difficult because the number of active devices, or traffic, changes throughout the day. This thesis explores whether computer programs that are trained and learn from experience can make better decisions about when to power base stations on or off. These programs learn by trying actions and receiving feedback on whether the choice was good or bad. Usually, this feedback is designed before the computer program's learning process and remains fixed throughout all the computer program's interactions. Our work demonstrates that a computer program trained using fixed feedback is ineffective when user demand fluctuates. We propose a two-level decision system in which one program selects, based on the user demand, which other computer programs will be active. Each of the eligible computer programs was trained with different feedback, therefore having different priorities regarding saving energy or maintaining service quality. This approach reduces energy use by up to 64% while keeping service performance at an acceptable level. These results suggest that learning-based decision systems can help future wireless networks operate more efficiently and sustainably. | en |
| dc.description.degree | Master of Science | en |
| dc.format.medium | ETD | en |
| dc.identifier.other | vt_gsexam:45243 | en |
| dc.identifier.uri | https://hdl.handle.net/10919/140536 | en |
| dc.language.iso | en | en |
| dc.publisher | Virginia Tech | en |
| dc.rights | In Copyright | en |
| dc.rights.uri | http://rightsstatements.org/vocab/InC/1.0/ | en |
| dc.subject | Energy efficiency | en |
| dc.subject | Base Station Switching | en |
| dc.subject | Hierarchical Reinforcement Learning | en |
| dc.title | From Static to Adaptive: Dynamic Cost Function Weight Adaptation in Hierarchical Reinforcement Learning for Sustainable 6G Radio Access Networks | en |
| dc.type | Thesis | en |
| thesis.degree.discipline | Computer Engineering | en |
| thesis.degree.grantor | Virginia Polytechnic Institute and State University | en |
| thesis.degree.level | masters | en |
| thesis.degree.name | Master of Science | en |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Viana_Fonseca_Abreu_J_T_2025.pdf
- Size:
- 511.28 KB
- Format:
- Adobe Portable Document Format