论文标题
一种使用量量化的联合加固学习方法,用于雾无线电访问网络中的合作边缘缓存
A Federated Reinforcement Learning Method with Quantization for Cooperative Edge Caching in Fog Radio Access Networks
论文作者
论文摘要
在本文中,在雾无线电网络(F-RAN)中研究了合作边缘缓存问题。鉴于问题的非确定性多项式硬(NP-HARD)属性,建议基于决斗的Q网络(决斗DQN)的缓存更新算法,以通过学习动态网络环境来做出最佳的缓存决定。为了保护用户数据隐私并解决单个深度强化学习(DRL)模型培训缓慢收敛的问题,我们建议使用量化的联合加固学习方法(FRLQ)从F-RAN中的多个FOG接入点(F-aps)实施模型的合作培训。为了解决由模型传输引起的通信资源的过度消耗,我们修剪和量化共享的DRL模型以减少模型传输参数的数量。通信间隔增加了,通过定期模型全球聚合减少了通信回合。我们分析了政策的全球融合和计算复杂性。仿真结果证明,与基准方案相比,我们的策略在减少用户请求延迟和提高缓存命中率方面具有更好的性能。提出的政策还显示出具有更快的训练速度和更高的通信效率,而模型准确性的损失最小。
In this paper, cooperative edge caching problem is studied in fog radio access networks (F-RANs). Given the non-deterministic polynomial hard (NP-hard) property of the problem, a dueling deep Q network (Dueling DQN) based caching update algorithm is proposed to make an optimal caching decision by learning the dynamic network environment. In order to protect user data privacy and solve the problem of slow convergence of the single deep reinforcement learning (DRL) model training, we propose a federated reinforcement learning method with quantization (FRLQ) to implement cooperative training of models from multiple fog access points (F-APs) in F-RANs. To address the excessive consumption of communications resources caused by model transmission, we prune and quantize the shared DRL models to reduce the number of model transfer parameters. The communications interval is increased and the communications rounds are reduced by periodical model global aggregation. We analyze the global convergence and computational complexity of our policy. Simulation results verify that our policy has better performance in reducing user request delay and improving cache hit rate compared to benchmark schemes. The proposed policy is also shown to have faster training speed and higher communications efficiency with minimal loss of model accuracy.