Please use this identifier to cite or link to this item: http://hdl.handle.net/123456789/2084
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSahasrabudhe, Neeraja-
dc.date.accessioned2020-11-24T04:30:23Z-
dc.date.available2020-11-24T04:30:23Z-
dc.date.issued2018-
dc.identifier.citationJournal of Machine Learning Research, 18, pp. 1-27en_US
dc.identifier.urihttps://jmlr.csail.mit.edu/papers/v18/15-592.html-
dc.identifier.urihttp://hdl.handle.net/123456789/2084-
dc.descriptionOnly IISERM authors are available in the record.-
dc.description.abstractWe propose a scheme for finding a “good” estimator for the gradient of a function on a high-dimensional space with few function evaluations, for applications where function evaluations are expensive and the function under consideration is not sensitive in all coordinates locally, making its gradient almost sparse. Exploiting the latter aspect, our method combines ideas from Spall’s Simultaneous Perturbation Stochastic Approximation with compressive sensing. We theoretically justify its computational advantages and illustrate them empirically by numerical experiments. In particular, applications to estimating gradient outer product matrix as well as standard optimization problems are illustrated via simulations.en_US
dc.language.isoenen_US
dc.publisherMicrotome Publishingen_US
dc.subjectGradient estimationen_US
dc.subjectCompressive sensingen_US
dc.subjectGradient descenten_US
dc.subjectGradient outer product matrix.en_US
dc.titleGradient estimation with simultaneous perturbation and compressive sensingen_US
dc.typeArticleen_US
Appears in Collections:Research Articles

Files in This Item:
File Description SizeFormat 
Need to add pdf.odt8.63 kBOpenDocument TextView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.