论文标题
DDPG驱动的深度解链,具有自适应深度,用于稀疏的贝叶斯学习频道估计
DDPG-Driven Deep-Unfolding with Adaptive Depth for Channel Estimation with Sparse Bayesian Learning
论文作者
论文摘要
深度无形的神经网络(NNS)受到了极大的关注,因为它们的复杂性相对较低。通常,这些深度折叠的NN仅限于所有输入的固定深度。但是,收敛所需的最佳图层随不同输入而变化。在本文中,我们首先开发了一个深层确定性策略梯度(DDPG)驱动的深度无折叠的框架,并针对不同输入进行自适应深度,在这些输入中,DDPG的深度折叠式NN的可训练参数是由DDPG学习的,而不是直接由随机梯度下降algorithm更新。具体而言,分别设计为DDPG的状态,动作和状态过渡,优化变量,可训练的参数和架构分别设计为已设计为状态,动作和状态过渡。然后,使用此框架来处理大量多输入多输出系统中的通道估计问题。具体而言,首先,我们以离网为基础制定了通道估计问题,并开发了基于稀疏的贝叶斯学习(SBL)算法来解决它。其次,将基于SBL的算法展开为一组带有一组可训练参数的层结构。第三,使用基于SBL的算法的不展开结构来解决此通道估计问题,采用了提出的DDPG驱动的深度解释框架。为了实现自适应深度,我们设计了停止分数以指示何时停止,这是通道重建误差的函数。此外,提出的框架被扩展到实现一般深度神经网络(DNNS)的自适应深度。仿真结果表明,所提出的算法的表现优于固定深度的常规优化算法和DNN,并且层数量降低。
Deep-unfolding neural networks (NNs) have received great attention since they achieve satisfactory performance with relatively low complexity. Typically, these deep-unfolding NNs are restricted to a fixed-depth for all inputs. However, the optimal number of layers required for convergence changes with different inputs. In this paper, we first develop a framework of deep deterministic policy gradient (DDPG)-driven deep-unfolding with adaptive depth for different inputs, where the trainable parameters of deep-unfolding NN are learned by DDPG, rather than updated by the stochastic gradient descent algorithm directly. Specifically, the optimization variables, trainable parameters, and architecture of deep-unfolding NN are designed as the state, action, and state transition of DDPG, respectively. Then, this framework is employed to deal with the channel estimation problem in massive multiple-input multiple-output systems. Specifically, first of all we formulate the channel estimation problem with an off-grid basis and develop a sparse Bayesian learning (SBL)-based algorithm to solve it. Secondly, the SBL-based algorithm is unfolded into a layer-wise structure with a set of introduced trainable parameters. Thirdly, the proposed DDPG-driven deep-unfolding framework is employed to solve this channel estimation problem based on the unfolded structure of the SBL-based algorithm. To realize adaptive depth, we design the halting score to indicate when to stop, which is a function of the channel reconstruction error. Furthermore, the proposed framework is extended to realize the adaptive depth of the general deep neural networks (DNNs). Simulation results show that the proposed algorithm outperforms the conventional optimization algorithms and DNNs with fixed depth with much reduced number of layers.