TY - GEN
T1 - A distributed reinforcement learning approach to maximize resource utilization and control handover dropping in multimedia wireless networks
AU - Alexandri, Eftychia
AU - Martinez, Gorges
AU - Zeghlache, Djamal
PY - 2002/12/1
Y1 - 2002/12/1
N2 - A new scheme to maximize resource utilization in a cellular network while respecting constraints on handover dropping probability is proposed and analyzed. The constraints are set for each traffic class separately and have to be respected by the network independently of the area in a localized manner. The problem is formulated as a Markov Decision Process (MDP) and solved by making use of the model-free simulation-based Q-learning algorithm that runs at each cell. Integration of the handover limit in the model is achieved by observing which of the new call arrivals, at a particular state of the system, are mostly responsible for violation of the handover dropping limit. Through trial and error, the algorithm proceeds to the statistical elimination of new admissions in the system, those causing excessive dropping. Results obtained via the proposed Reinforcement Learning (RL) based approach are compared with a resource allocation that takes into consideration heterogeneous and unevenly distributed traffic over the geographical area under consideration. For the scenarios examined, comparable results and performance are observed with an advantage for RL in blocking and utilization.
AB - A new scheme to maximize resource utilization in a cellular network while respecting constraints on handover dropping probability is proposed and analyzed. The constraints are set for each traffic class separately and have to be respected by the network independently of the area in a localized manner. The problem is formulated as a Markov Decision Process (MDP) and solved by making use of the model-free simulation-based Q-learning algorithm that runs at each cell. Integration of the handover limit in the model is achieved by observing which of the new call arrivals, at a particular state of the system, are mostly responsible for violation of the handover dropping limit. Through trial and error, the algorithm proceeds to the statistical elimination of new admissions in the system, those causing excessive dropping. Results obtained via the proposed Reinforcement Learning (RL) based approach are compared with a resource allocation that takes into consideration heterogeneous and unevenly distributed traffic over the geographical area under consideration. For the scenarios examined, comparable results and performance are observed with an advantage for RL in blocking and utilization.
UR - https://www.scopus.com/pages/publications/44849138630
U2 - 10.1109/PIMRC.2002.1046544
DO - 10.1109/PIMRC.2002.1046544
M3 - Conference contribution
AN - SCOPUS:44849138630
SN - 0780375890
SN - 9780780375895
T3 - IEEE International Symposium on Personal, Indoor and Mobile Radio Communications, PIMRC
SP - 2249
EP - 2253
BT - 13th IEEE International Symposium on Personal, Indoor and Mobile Radio Communications, PIMRC 2002
T2 - 13th IEEE International Symposium on Personal, Indoor and Mobile Radio Communications, PIMRC 2002
Y2 - 15 September 2002 through 18 September 2002
ER -