Offline Reinforcement Learning for Mobility Robustness Optimization
Konferenz: European WIRELESS 2025 - 30th European Wireless Conference
27.10.2025-29.10.2025 in Sohia Antipolis, France
Tagungsband: European Wireless 2025
Seiten: 7Sprache: EnglischTyp: PDF
Autoren:
Alizadeh, Pegah; Giovanidis, Anastasios; Ramachandra, Pradeepa; Koutsoukis, Vasileios; Arouk, Osama
Inhalt:
In this work we revisit the Mobility Robustness Optimization (MRO) algorithm and study the possibility of learning the optimal Cell Individual Offset tuning using offline Reinforcement Learning. Such methods make use of collected offline datasets to learn the optimal policy, without further exploration. We adapt and apply a sequence-based method called Decision Transformers as well as a value-based method called Conservative Q-Learning to learn the optimal policy for the same target reward as the vanilla rule-based MRO. The same input features related to failures, ping-pongs, and other handover issues are used. Evaluation for realistic New Radio networks with 3500 MHz carrier frequency on a traffic mix including diverse user service types and a specific tunable cell-pair shows that offline- RL methods outperform rule-based MRO, offering up to 7% improvement. Furthermore, offline-RL can be trained for diverse objective functions using the same available dataset, thus offering operational flexibility compared to rule-based methods.

