File Download
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1137/22M1510261
- Scopus: eid_2-s2.0-85200122759
- WOS: WOS:001343301000009
- Find via

Supplementary
- Citations:
- Appears in Collections:
Article: Small Errors in Random Zeroth-Order Optimization Are Imaginary
| Title | Small Errors in Random Zeroth-Order Optimization Are Imaginary |
|---|---|
| Authors | |
| Keywords | complex-step derivative derivative-free optimization zeroth-order optimization |
| Issue Date | 19-Jul-2024 |
| Publisher | Society for Industrial and Applied Mathematics |
| Citation | SIAM Journal on Optimization, 2024, v. 34, n. 3, p. 2638-2670 How to Cite? |
| Abstract | Most zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some gradient estimator that can be computed from a small number of function evaluations. This estimator is constructed randomly, and its expectation matches the gradient of a smooth approximation of the objective function whose quality improves as the underlying smoothing parameter δ is reduced. Gradient estimators requiring a smaller number of function evaluations are preferable from a computational point of view. While estimators based on a single function evaluation can be obtained by use of the divergence theorem from vector calculus, their variance explodes as δ tends to 0. Estimators based on multiple function evaluations, on the other hand, suffer from numerical cancellation when δ tends to 0. To combat both effects simultaneously, we extend the objective function to the complex domain and construct a gradient estimator that evaluates the objective at a complex point whose coordinates have small imaginary parts of the order δ . As this estimator requires only one function evaluation, it is immune to cancellation. In addition, its variance remains bounded as δ tends to 0. We prove that zeroth-order algorithms that use our estimator offer the same theoretical convergence guarantees as the state-of-the-art methods. Numerical experiments suggest, however, that they often converge faster in practice. |
| Persistent Identifier | http://hdl.handle.net/10722/346072 |
| ISSN | 2023 Impact Factor: 2.6 2023 SCImago Journal Rankings: 2.138 |
| ISI Accession Number ID |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Jongeneel, Wouter | - |
| dc.contributor.author | Yue, Man Chung | - |
| dc.contributor.author | Kuhn, Daniel | - |
| dc.date.accessioned | 2024-09-07T00:30:27Z | - |
| dc.date.available | 2024-09-07T00:30:27Z | - |
| dc.date.issued | 2024-07-19 | - |
| dc.identifier.citation | SIAM Journal on Optimization, 2024, v. 34, n. 3, p. 2638-2670 | - |
| dc.identifier.issn | 1052-6234 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/346072 | - |
| dc.description.abstract | Most zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some gradient estimator that can be computed from a small number of function evaluations. This estimator is constructed randomly, and its expectation matches the gradient of a smooth approximation of the objective function whose quality improves as the underlying smoothing parameter δ is reduced. Gradient estimators requiring a smaller number of function evaluations are preferable from a computational point of view. While estimators based on a single function evaluation can be obtained by use of the divergence theorem from vector calculus, their variance explodes as δ tends to 0. Estimators based on multiple function evaluations, on the other hand, suffer from numerical cancellation when δ tends to 0. To combat both effects simultaneously, we extend the objective function to the complex domain and construct a gradient estimator that evaluates the objective at a complex point whose coordinates have small imaginary parts of the order δ . As this estimator requires only one function evaluation, it is immune to cancellation. In addition, its variance remains bounded as δ tends to 0. We prove that zeroth-order algorithms that use our estimator offer the same theoretical convergence guarantees as the state-of-the-art methods. Numerical experiments suggest, however, that they often converge faster in practice. | - |
| dc.language | eng | - |
| dc.publisher | Society for Industrial and Applied Mathematics | - |
| dc.relation.ispartof | SIAM Journal on Optimization | - |
| dc.subject | complex-step derivative | - |
| dc.subject | derivative-free optimization | - |
| dc.subject | zeroth-order optimization | - |
| dc.title | Small Errors in Random Zeroth-Order Optimization Are Imaginary | - |
| dc.type | Article | - |
| dc.description.nature | published_or_final_version | - |
| dc.identifier.doi | 10.1137/22M1510261 | - |
| dc.identifier.scopus | eid_2-s2.0-85200122759 | - |
| dc.identifier.volume | 34 | - |
| dc.identifier.issue | 3 | - |
| dc.identifier.spage | 2638 | - |
| dc.identifier.epage | 2670 | - |
| dc.identifier.eissn | 1095-7189 | - |
| dc.identifier.isi | WOS:001343301000009 | - |
| dc.identifier.issnl | 1052-6234 | - |
