Abstract
The integration of subterranean LoRaWAN and non-terrestrial networks (NTN) delivers substantial economic and societal benefits in remote agriculture and disaster rescue operations. The LoRa modulation leverages quasi-orthogonal spreading factors (SFs) to optimize data rates, airtime, coverage and energy consumption. However, it is still challenging to effectively assign SFs to end devices for minimizing co-SF interference in massive subterranean LoRaWAN NTN. To address this, we investigate a reinforcement learning (RL)-based SFs allocation scheme to optimize the system’s energy efficiency (EE). To efficiently capture the device-to-environment interactions in dense networks, we proposed an SFs allocation technique using the multi-agent dueling double deep Q-network (MAD3QN) and the multi-agent advantage actor-critic (MAA2C) algorithms based on an analytical reward mechanism. Our proposed RL-based SFs allocation approach evinces better performance compared to four benchmarks in the extreme underground direct-to-satellite scenario. Remarkably, MAD3QN shows promising potentials in surpassing MAA2C in terms of convergence rate and EE.
Original language | English |
---|---|
Pages (from-to) | 308-312 |
Number of pages | 5 |
Journal | IEEE Wireless Communications Letters |
Volume | 13 |
Issue number | 2 |
Early online date | 26 Oct 2023 |
DOIs | |
Publication status | Published - 1 Feb 2024 |
MoE publication type | A1 Journal article-refereed |
Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 42211530077 and Grant 42074179; in part by the Academy of Finland, 6G Flagship Program under Grant 346208; and in part by the China Scholarship Council.
Keywords
- SFs allocation
- non-terrestrial networks
- energy efficiency
- Interference
- Reinforcement learning
- Soil
- Energy efficiency
- Resource management
- Uplink
- Subterranean LoRaWAN
- reinforcement learning
- Signal to noise ratio