DS1 spectrogram: Classical and Quantum Speedups for Non-Convex Optimization via Energy Conserving Descent

Classical and Quantum Speedups for Non-Convex Optimization via Energy Conserving Descent

April 14, 20262604.13022

Authors

Yihang Sun,Huaijin Wang,Patrick Hayden,Jose Blanchet

Abstract

The Energy Conserving Descent (ECD) algorithm was recently proposed (De Luca & Silverstein, 2022) as a global non-convex optimization method. Unlike gradient descent, appropriately configured ECD dynamics escape strict local minima and converge to a global minimum, making it appealing for machine learning optimization.

We present the first analytical study of ECD, focusing on the one-dimensional setting for this first installment. We formalize a stochastic ECD dynamics (sECD) with energy-preserving noise, as well as a quantum analog of the ECD Hamiltonian (qECD), providing the foundation for a quantum algorithm through Hamiltonian simulation.

For positive double-well objectives, we compute the expected hitting time from a local to the global minimum. We prove that both sECD and qECD yield exponential speedup over respective gradient descent baselines--stochastic gradient descent and its quantization.

For objectives with tall barriers, qECD achieves a further speedup over sECD.

Resources

Stay in the loop

Every AI paper that matters, free in your inbox daily.

Details

  • © 2026 takara.ai Ltd
  • Content is sourced from third-party publications.