论文标题
具有继电器的物联网网络中的实用AOI调度程序
A Practical AoI Scheduler in IoT Networks with Relays
论文作者
论文摘要
随着自主计算,设备之间的通信和协作在完成各种任务方面,物联网(IoT)网络已变得无处不在。物联网网络中的继电器进一步使部署IoT网络变得方便,因为继电器提供了许多好处,例如增加通信范围并最大程度地减少功耗。现有有关此类两跳的物联网网络的传统AOI调度程序的文献受到限制,因为它们的设计是假设恒定/非变化的频道条件,并且已知(通常是在意外生成)数据包生成模式。已经对具有继电器的两跳物IoT网络中的AOI调度进行了深入的加固学习(DRL)算法,但是,由于网络变得较大,它们仅适用于小规模的IoT网络。这些限制阻止了对IoT网络部署的AOI调度程序的实际利用。本文为两跳物IoT网络提供了一个实用的AOI调度程序,并提供了解决上述限制的继电器。拟议的调度程序利用了一种基于新型投票机制的近端策略优化(V-PPO)算法,该算法维持线性动作空间,从而可以通过较大的物联网网络进行良好的比例。拟议的基于V-PPO的AOI调度程序很好地适应了不知所措的网络条件,并说明了未知的流量生成模式,这使其对于现实世界的IoT部署而实用。仿真结果表明,提出的基于V-PPO的AOI调度程序优于ML和传统(非ML)AOI调度程序,例如,基于DEED Q Network(DQN)的AOI调度程序,最大年龄的最大年龄最大年龄差异年龄差异(MAF-MAD),MAF-MAD(MAF-MAD),MAF(Mafimal Age First)和所有考虑练习场景。
Internet of Things (IoT) networks have become ubiquitous as autonomous computing, communication and collaboration among devices become popular for accomplishing various tasks. The use of relays in IoT networks further makes it convenient to deploy IoT networks as relays provide a host of benefits, like increasing the communication range and minimizing power consumption. Existing literature on traditional AoI schedulers for such two-hop relayed IoT networks are limited because they are designed assuming constant/non-changing channel conditions and known (usually, generate-at-will) packet generation patterns. Deep reinforcement learning (DRL) algorithms have been investigated for AoI scheduling in two-hop IoT networks with relays, however, they are only applicable for small-scale IoT networks due to exponential rise in action space as the networks become large. These limitations discourage the practical utilization of AoI schedulers for IoT network deployments. This paper presents a practical AoI scheduler for two-hop IoT networks with relays that addresses the above limitations. The proposed scheduler utilizes a novel voting mechanism based proximal policy optimization (v-PPO) algorithm that maintains a linear action space, enabling it be scale well with larger IoT networks. The proposed v-PPO based AoI scheduler adapts well to changing network conditions and accounts for unknown traffic generation patterns, making it practical for real-world IoT deployments. Simulation results show that the proposed v-PPO based AoI scheduler outperforms both ML and traditional (non-ML) AoI schedulers, such as, Deep Q Network (DQN)-based AoI Scheduler, Maximal Age First-Maximal Age Difference (MAF-MAD), MAF (Maximal Age First) , and round-robin in all considered practical scenarios.