La Trobe

Constrained Reinforcement Learning for Resource Allocation in Network Slicing

Download (554.42 kB)
journal contribution
posted on 2025-02-11, 04:10 authored by Yizhen Xu, Zhengyang Zhao, Peng ChengPeng Cheng, Zhuo Chen, Ming Ding, Branka Vucetic, Yonghui Li
In network slicing, dynamic resource allocation is the key to network performance optimization. Deep reinforcement learning (DRL) is a promising method to exploit the dynamic features of network slicing by interacting with the environment. However, the existing DRL-based resource allocation solutions can only handle a discrete action space. In this letter, we tackle a general DRL-based resource allocation problem which considers a mixed action space including both discrete channel allocation and continuous energy harvesting time division, with the constraints of energy consumption and queue package length. We propose a novel DRL algorithm referred to as constrained discrete-continuous soft actor-critic (CDC-SAC) by redesigning the network architecture and policy learning process. Simulation results show that the proposed algorithm can achieve a significant performance improvement in terms of the total throughput with the strict constraints guarantee.

Funding

The work of Peng Cheng was supported by ARC under Grant DE190100162 and DP210103410. The work of Yonghui Li was supported by ARC under Grant DP190101988 and DP210103410.

History

Publication Date

2021-05-01

Journal

IEEE Communications Letters

Volume

25

Issue

5

Pagination

1554-1558

Publisher

Institute of Electrical and Electronics Engineers

ISSN

1089-7798

Rights Statement

© 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.