First Order Online Optimisation Using Forward Gradients in Over-Parameterised Systems

Behnam Mafakheri, Jonathan H. Manton, Iman Shames

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

The success of deep learning over the past decade mainly relies on gradient-based optimisation and backpropagation. This paper focuses on analysing the performance of first-order gradient-based optimisation algorithms with time-varying non-convex cost function under Polyak-Lojasiewicz condition. Specifically, we focus on using the forward mode of automatic differentiation to compute directional derivatives of the cost function in fast-changing problems where calculating gradients using the backpropagation algorithm is either impossible or inefficient. Upper bounds for tracking and asymptotic errors are derived for various cases, showing the linear convergence to a solution or a neighbourhood of an optimal solution, where the convergence rate decreases with the increase in the dimension of the problem. We present numerical results demonstrating the method's correctness and performance.

Original languageEnglish
Title of host publication2024 European Control Conference, ECC 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages2119-2124
Number of pages6
ISBN (Electronic)9783907144107
DOIs
Publication statusPublished - 2024
Event2024 European Control Conference, ECC 2024 - Stockholm, Sweden
Duration: 25 Jun 202428 Jun 2024

Publication series

Name2024 European Control Conference, ECC 2024

Conference

Conference2024 European Control Conference, ECC 2024
Country/TerritorySweden
CityStockholm
Period25/06/2428/06/24

Fingerprint

Dive into the research topics of 'First Order Online Optimisation Using Forward Gradients in Over-Parameterised Systems'. Together they form a unique fingerprint.

Cite this