TY - JOUR
T1 - A Deep Reinforcement Learning-Based Dynamic Traffic Offloading in Space-Air-Ground Integrated Networks (SAGIN)
AU - Tang, Fengxiao
AU - Hofner, Hans
AU - Kato, Nei
AU - Kaneko, Kazuma
AU - Yamashita, Yasutaka
AU - Hangai, Masatake
N1 - Publisher Copyright:
© 1983-2012 IEEE.
PY - 2022/1/1
Y1 - 2022/1/1
N2 - Space-Air-Ground Integrated Networks (SAGIN) is considered as the key structure of the next generation network. The space satellites and air nodes are the potential candidates to assist and offload the terrain transmissions. However, due to the high mobility of space and air nodes as well as the high dynamic of network traffic, the conventional traffic offloading strategy is not applicable for the high dynamic SAGIN. In this paper, we propose a reinforcement learning based traffic offloading for SAGIN by considering the high mobility of nodes as well as frequent changing network traffic and link state. In the proposal, a double Q-learning algorithm with improved delay-sensitive replay memory algorithm (DSRPM) is proposed to train the node to decide offloading strategy based on the local and neighboring historical information. Furthermore, a joint information collection with hello package and offline training mechanism is proposed to assist the proposed offloading algorithm. The simulation shows that the proposal outperforms conventional offloading algorithms in terms of signaling overhead, dynamic adaptivity, packet drop rate and transmission delay.
AB - Space-Air-Ground Integrated Networks (SAGIN) is considered as the key structure of the next generation network. The space satellites and air nodes are the potential candidates to assist and offload the terrain transmissions. However, due to the high mobility of space and air nodes as well as the high dynamic of network traffic, the conventional traffic offloading strategy is not applicable for the high dynamic SAGIN. In this paper, we propose a reinforcement learning based traffic offloading for SAGIN by considering the high mobility of nodes as well as frequent changing network traffic and link state. In the proposal, a double Q-learning algorithm with improved delay-sensitive replay memory algorithm (DSRPM) is proposed to train the node to decide offloading strategy based on the local and neighboring historical information. Furthermore, a joint information collection with hello package and offline training mechanism is proposed to assist the proposed offloading algorithm. The simulation shows that the proposal outperforms conventional offloading algorithms in terms of signaling overhead, dynamic adaptivity, packet drop rate and transmission delay.
KW - Double Q-learning
KW - Reinforcement learning (RL)
KW - Satellite communication
KW - Space-air-ground integrated networks (SAGIN)
KW - Traffic offloading
KW - UAV
UR - http://www.scopus.com/inward/record.url?scp=85121878773&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85121878773&partnerID=8YFLogxK
U2 - 10.1109/JSAC.2021.3126073
DO - 10.1109/JSAC.2021.3126073
M3 - Article
AN - SCOPUS:85121878773
SN - 0733-8716
VL - 40
SP - 276
EP - 289
JO - IEEE Journal on Selected Areas in Communications
JF - IEEE Journal on Selected Areas in Communications
IS - 1
ER -