File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Small Errors in Random Zeroth-Order Optimization Are Imaginary

TitleSmall Errors in Random Zeroth-Order Optimization Are Imaginary
Authors
Keywordscomplex-step derivative
derivative-free optimization
zeroth-order optimization
Issue Date19-Jul-2024
PublisherSociety for Industrial and Applied Mathematics
Citation
SIAM Journal on Optimization, 2024, v. 34, n. 3, p. 2638-2670 How to Cite?
AbstractMost zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some gradient estimator that can be computed from a small number of function evaluations. This estimator is constructed randomly, and its expectation matches the gradient of a smooth approximation of the objective function whose quality improves as the underlying smoothing parameter δ is reduced. Gradient estimators requiring a smaller number of function evaluations are preferable from a computational point of view. While estimators based on a single function evaluation can be obtained by use of the divergence theorem from vector calculus, their variance explodes as δ tends to 0. Estimators based on multiple function evaluations, on the other hand, suffer from numerical cancellation when δ tends to 0. To combat both effects simultaneously, we extend the objective function to the complex domain and construct a gradient estimator that evaluates the objective at a complex point whose coordinates have small imaginary parts of the order δ . As this estimator requires only one function evaluation, it is immune to cancellation. In addition, its variance remains bounded as δ tends to 0. We prove that zeroth-order algorithms that use our estimator offer the same theoretical convergence guarantees as the state-of-the-art methods. Numerical experiments suggest, however, that they often converge faster in practice.
Persistent Identifierhttp://hdl.handle.net/10722/346072
ISSN
2023 Impact Factor: 2.6
2023 SCImago Journal Rankings: 2.138
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorJongeneel, Wouter-
dc.contributor.authorYue, Man Chung-
dc.contributor.authorKuhn, Daniel-
dc.date.accessioned2024-09-07T00:30:27Z-
dc.date.available2024-09-07T00:30:27Z-
dc.date.issued2024-07-19-
dc.identifier.citationSIAM Journal on Optimization, 2024, v. 34, n. 3, p. 2638-2670-
dc.identifier.issn1052-6234-
dc.identifier.urihttp://hdl.handle.net/10722/346072-
dc.description.abstractMost zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some gradient estimator that can be computed from a small number of function evaluations. This estimator is constructed randomly, and its expectation matches the gradient of a smooth approximation of the objective function whose quality improves as the underlying smoothing parameter δ is reduced. Gradient estimators requiring a smaller number of function evaluations are preferable from a computational point of view. While estimators based on a single function evaluation can be obtained by use of the divergence theorem from vector calculus, their variance explodes as δ tends to 0. Estimators based on multiple function evaluations, on the other hand, suffer from numerical cancellation when δ tends to 0. To combat both effects simultaneously, we extend the objective function to the complex domain and construct a gradient estimator that evaluates the objective at a complex point whose coordinates have small imaginary parts of the order δ . As this estimator requires only one function evaluation, it is immune to cancellation. In addition, its variance remains bounded as δ tends to 0. We prove that zeroth-order algorithms that use our estimator offer the same theoretical convergence guarantees as the state-of-the-art methods. Numerical experiments suggest, however, that they often converge faster in practice.-
dc.languageeng-
dc.publisherSociety for Industrial and Applied Mathematics-
dc.relation.ispartofSIAM Journal on Optimization-
dc.subjectcomplex-step derivative-
dc.subjectderivative-free optimization-
dc.subjectzeroth-order optimization-
dc.titleSmall Errors in Random Zeroth-Order Optimization Are Imaginary-
dc.typeArticle-
dc.description.naturepublished_or_final_version-
dc.identifier.doi10.1137/22M1510261-
dc.identifier.scopuseid_2-s2.0-85200122759-
dc.identifier.volume34-
dc.identifier.issue3-
dc.identifier.spage2638-
dc.identifier.epage2670-
dc.identifier.eissn1095-7189-
dc.identifier.isiWOS:001343301000009-
dc.identifier.issnl1052-6234-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats