The growing number of Internet of Things (IoT) devices has increased demand for wireless spectrum, exacerbating the spectrum scarcity issue. While dynamic spectrum access (DSA) has emerged as a promising solution, most existing works focus on channel selection, neglecting other control factors. To address this, we develop a deep reinforcement learning (DRL) environment that involves both spectrum access and power control, and propose a dual-head output DRL structure that efficiently handles this mixed action space in a unified model. Moreover, we represent Recurrent Flash (ReF) framework, which integrates linear attention with a recurrent mechanism to enhance the temporal memory abilities of agents. Utilizing ReF as representation learning layers of DRL approaches enables a more precise comprehension of environment dynamics and promotes better-informed decision-making. The combination of the dual-head output and ReF, referred to as ReF-RL, is rigorously evaluated through experiments. Results demonstrate that ReF-RL significantly improves communication quality and device longevity, thereby outperforming both conventional communication and existing DRL algorithms.

Recurrent Flash Reinforcement Learning for Dynamic Spectrum Access and Power Control

Pace P.;Fortino G.
2024-01-01

Abstract

The growing number of Internet of Things (IoT) devices has increased demand for wireless spectrum, exacerbating the spectrum scarcity issue. While dynamic spectrum access (DSA) has emerged as a promising solution, most existing works focus on channel selection, neglecting other control factors. To address this, we develop a deep reinforcement learning (DRL) environment that involves both spectrum access and power control, and propose a dual-head output DRL structure that efficiently handles this mixed action space in a unified model. Moreover, we represent Recurrent Flash (ReF) framework, which integrates linear attention with a recurrent mechanism to enhance the temporal memory abilities of agents. Utilizing ReF as representation learning layers of DRL approaches enables a more precise comprehension of environment dynamics and promotes better-informed decision-making. The combination of the dual-head output and ReF, referred to as ReF-RL, is rigorously evaluated through experiments. Results demonstrate that ReF-RL significantly improves communication quality and device longevity, thereby outperforming both conventional communication and existing DRL algorithms.
2024
Deep Reinforcement Learning
Dynamic Spectrum Access
Linear Attention
Recurrent Flash Framework
Resource Allocation Optimization
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.11770/379426
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact