论文标题
基于策略梯度的量子近似优化算法
Policy Gradient based Quantum Approximate Optimization Algorithm
论文作者
论文摘要
最近,量子近似优化算法(QAOA)作为一种杂种量子/经典算法,最近引起了很大的兴趣。 QAOA也可以看作是用于量子控制的变异ansatz。但是,其直接应用于新兴的量子技术会遇到其他物理约束:(i)量子系统的状态无法观察到; (ii)获得目标函数的衍生物在实验中可能是昂贵甚至无法访问的,并且(iii)目标函数的值可能对各种不确定性来源敏感,就像嘈杂的中等规模量子(NISQ)设备的情况一样。考虑到这种约束,我们表明基于策略阶段的增强算法(RL)算法非常适合以噪声方式优化QAOA的变异参数,为开发连续量子控制的RL技术开辟了道路。这有助于减轻和监视现代量子模拟器中可能未知的错误来源。我们分析了单个和多量系统中量子状态转移问题的算法的性能,但要符合各种噪声来源,例如哈密顿量中的误差项或测量过程中的量子不确定性。我们表明,在嘈杂的设置中,它能够超过现有的优化算法的表现。
The quantum approximate optimization algorithm (QAOA), as a hybrid quantum/classical algorithm, has received much interest recently. QAOA can also be viewed as a variational ansatz for quantum control. However, its direct application to emergent quantum technology encounters additional physical constraints: (i) the states of the quantum system are not observable; (ii) obtaining the derivatives of the objective function can be computationally expensive or even inaccessible in experiments, and (iii) the values of the objective function may be sensitive to various sources of uncertainty, as is the case for noisy intermediate-scale quantum (NISQ) devices. Taking such constraints into account, we show that policy-gradient-based reinforcement learning (RL) algorithms are well suited for optimizing the variational parameters of QAOA in a noise-robust fashion, opening up the way for developing RL techniques for continuous quantum control. This is advantageous to help mitigate and monitor the potentially unknown sources of errors in modern quantum simulators. We analyze the performance of the algorithm for quantum state transfer problems in single- and multi-qubit systems, subject to various sources of noise such as error terms in the Hamiltonian, or quantum uncertainty in the measurement process. We show that, in noisy setups, it is capable of outperforming state-of-the-art existing optimization algorithms.