Offline Reinforcement Learning for Mobility Robustness Optimization
Conference: European WIRELESS 2025 - 30th European Wireless Conference
10/27/2025 - 10/29/2025 at Sohia Antipolis, France
Proceedings: European Wireless 2025
Pages: 7Language: englishTyp: PDF
Authors:
Alizadeh, Pegah; Giovanidis, Anastasios; Ramachandra, Pradeepa; Koutsoukis, Vasileios; Arouk, Osama
Abstract:
In this work we revisit the Mobility Robustness Optimization (MRO) algorithm and study the possibility of learning the optimal Cell Individual Offset tuning using offline Reinforcement Learning. Such methods make use of collected offline datasets to learn the optimal policy, without further exploration. We adapt and apply a sequence-based method called Decision Transformers as well as a value-based method called Conservative Q-Learning to learn the optimal policy for the same target reward as the vanilla rule-based MRO. The same input features related to failures, ping-pongs, and other handover issues are used. Evaluation for realistic New Radio networks with 3500 MHz carrier frequency on a traffic mix including diverse user service types and a specific tunable cell-pair shows that offline- RL methods outperform rule-based MRO, offering up to 7% improvement. Furthermore, offline-RL can be trained for diverse objective functions using the same available dataset, thus offering operational flexibility compared to rule-based methods.

