TY - GEN
T1 - Advancing RAN Slicing with Offline Reinforcement Learning
AU - Yang, Kun
AU - Yeh, Shu Ping
AU - Zhang, Menglei
AU - Sydir, Jerry
AU - Yang, Jing
AU - Shen, Cong
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Dynamic radio resource management (RRM) in wireless networks presents significant challenges, particularly in the context of Radio Access Network (RAN) slicing. This technology, crucial for catering to varying user requirements, often grapples with complex optimization scenarios. Existing Reinforcement Learning (RL) approaches, while achieving good performance in RAN slicing, typically rely on online algorithms or behavior cloning. These methods necessitate either continuous environmental interactions or access to high-quality datasets, hindering their practical deployment. Towards addressing these limitations, this paper introduces offline RL to solving the RAN slicing problem, marking a significant shift toward more feasible and adaptive RRM methods. We demonstrate how offline RL can effectively learn near-optimal policies from sub-optimal datasets, a notable advancement over existing practices. Our research highlights the inherent flexibility of offline RL, showcasing its ability to adjust policy criteria without the need for additional environmental interactions. Furthermore, we present empirical evidence of the efficacy of offline RL in adapting to various service-level requirements, illustrating its potential in diverse RAN slicing scenarios.
AB - Dynamic radio resource management (RRM) in wireless networks presents significant challenges, particularly in the context of Radio Access Network (RAN) slicing. This technology, crucial for catering to varying user requirements, often grapples with complex optimization scenarios. Existing Reinforcement Learning (RL) approaches, while achieving good performance in RAN slicing, typically rely on online algorithms or behavior cloning. These methods necessitate either continuous environmental interactions or access to high-quality datasets, hindering their practical deployment. Towards addressing these limitations, this paper introduces offline RL to solving the RAN slicing problem, marking a significant shift toward more feasible and adaptive RRM methods. We demonstrate how offline RL can effectively learn near-optimal policies from sub-optimal datasets, a notable advancement over existing practices. Our research highlights the inherent flexibility of offline RL, showcasing its ability to adjust policy criteria without the need for additional environmental interactions. Furthermore, we present empirical evidence of the efficacy of offline RL in adapting to various service-level requirements, illustrating its potential in diverse RAN slicing scenarios.
UR - http://www.scopus.com/inward/record.url?scp=85202993048&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85202993048&partnerID=8YFLogxK
U2 - 10.1109/DySPAN60163.2024.10632750
DO - 10.1109/DySPAN60163.2024.10632750
M3 - Conference contribution
AN - SCOPUS:85202993048
T3 - 2024 IEEE International Symposium on Dynamic Spectrum Access Networks, DySPAN 2024
SP - 331
EP - 338
BT - 2024 IEEE International Symposium on Dynamic Spectrum Access Networks, DySPAN 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 IEEE International Symposium on Dynamic Spectrum Access Networks, DySPAN 2024
Y2 - 13 May 2024 through 16 May 2024
ER -