Inexact SARAH Algorithm for Stochastic Optimization
We develop and analyze a variant of the SARAH algorithm, which does not require computation of the exact gradient. Thus this new method can be applied to general expectation minimization problems rather than only finite sum problems. While the original SARAH algorithm, as well as its predecessor, SV...
Saved in:
| Main Authors | , , |
|---|---|
| Format | Journal Article |
| Language | English |
| Published |
25.11.2018
|
| Subjects | |
| Online Access | Get full text |
| DOI | 10.48550/arxiv.1811.10105 |
Cover
| Abstract | We develop and analyze a variant of the SARAH algorithm, which does not
require computation of the exact gradient. Thus this new method can be applied
to general expectation minimization problems rather than only finite sum
problems. While the original SARAH algorithm, as well as its predecessor, SVRG,
require an exact gradient computation on each outer iteration, the inexact
variant of SARAH (iSARAH), which we develop here, requires only stochastic
gradient computed on a mini-batch of sufficient size. The proposed method
combines variance reduction via sample size selection and iterative stochastic
gradient updates. We analyze the convergence rate of the algorithms for
strongly convex and non-strongly convex cases, under smooth assumption with
appropriate mini-batch size selected for each case. We show that with an
additional, reasonable, assumption iSARAH achieves the best known complexity
among stochastic methods in the case of non-strongly convex stochastic
functions. |
|---|---|
| AbstractList | We develop and analyze a variant of the SARAH algorithm, which does not
require computation of the exact gradient. Thus this new method can be applied
to general expectation minimization problems rather than only finite sum
problems. While the original SARAH algorithm, as well as its predecessor, SVRG,
require an exact gradient computation on each outer iteration, the inexact
variant of SARAH (iSARAH), which we develop here, requires only stochastic
gradient computed on a mini-batch of sufficient size. The proposed method
combines variance reduction via sample size selection and iterative stochastic
gradient updates. We analyze the convergence rate of the algorithms for
strongly convex and non-strongly convex cases, under smooth assumption with
appropriate mini-batch size selected for each case. We show that with an
additional, reasonable, assumption iSARAH achieves the best known complexity
among stochastic methods in the case of non-strongly convex stochastic
functions. |
| Author | Nguyen, Lam M Scheinberg, Katya Takáč, Martin |
| Author_xml | – sequence: 1 givenname: Lam M surname: Nguyen fullname: Nguyen, Lam M – sequence: 2 givenname: Katya surname: Scheinberg fullname: Scheinberg, Katya – sequence: 3 givenname: Martin surname: Takáč fullname: Takáč, Martin |
| BackLink | https://doi.org/10.48550/arXiv.1811.10105$$DView paper in arXiv |
| BookMark | eNrjYmDJy89LZWCQNDTQM7EwNTXQTyyqyCzTM7QwNNQzNDA0MOVkMPbMS61ITC5RCHYMcvRQcMxJzy_KLMnIVUjLL1IILslPzkgsLslMVvAvKMnMzaxKLMnMz-NhYE1LzClO5YXS3Azybq4hzh66YOPjC4oycxOLKuNB1sSDrTEmrAIA4ywzQA |
| ContentType | Journal Article |
| Copyright | http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
| Copyright_xml | – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
| DBID | AKY AKZ GOX |
| DOI | 10.48550/arxiv.1811.10105 |
| DatabaseName | arXiv Computer Science arXiv Mathematics arXiv.org |
| DatabaseTitleList | |
| Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
| DeliveryMethod | fulltext_linktorsrc |
| ExternalDocumentID | 1811_10105 |
| GroupedDBID | AKY AKZ GOX |
| ID | FETCH-arxiv_primary_1811_101053 |
| IEDL.DBID | GOX |
| IngestDate | Tue Jul 22 23:02:37 EDT 2025 |
| IsDoiOpenAccess | true |
| IsOpenAccess | true |
| IsPeerReviewed | false |
| IsScholarly | false |
| Language | English |
| LinkModel | DirectLink |
| MergedId | FETCHMERGED-arxiv_primary_1811_101053 |
| OpenAccessLink | https://arxiv.org/abs/1811.10105 |
| ParticipantIDs | arxiv_primary_1811_10105 |
| PublicationCentury | 2000 |
| PublicationDate | 2018-11-25 |
| PublicationDateYYYYMMDD | 2018-11-25 |
| PublicationDate_xml | – month: 11 year: 2018 text: 2018-11-25 day: 25 |
| PublicationDecade | 2010 |
| PublicationYear | 2018 |
| Score | 3.350194 |
| SecondaryResourceType | preprint |
| Snippet | We develop and analyze a variant of the SARAH algorithm, which does not
require computation of the exact gradient. Thus this new method can be applied
to... |
| SourceID | arxiv |
| SourceType | Open Access Repository |
| SubjectTerms | Computer Science - Learning Mathematics - Optimization and Control |
| Title | Inexact SARAH Algorithm for Stochastic Optimization |
| URI | https://arxiv.org/abs/1811.10105 |
| hasFullText | 1 |
| inHoldings | 1 |
| isFullTextHit | |
| isPrint | |
| link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdVxBS8MwGP2YPe0iyhxVp-bgtYw2ycyOZViqoAWn0FtJ0nQV3Cazk_58v6QVveyahOQj4eO9R3gP4BYRSynEiaDkhgeMSoo9V7JASC1oNeORptY7_PQ8S9_YY87zAZBfL4zcte_fXT6w-poi_IRWX9qQ0iMkCtbMm-Xd56SL4urX_61DjumG_oFEcgLHPbsjcfccpzAwmxHQh41ppW7IMn6JUxJ_rLaoyOs1Qb5Ils1W19KGJZMMu3fd2yLP4Ca5f12kgTum-OwyIQpbQeEqoGPwULkbHwhDuXAnDBc61GweaVWGjFeKVnJONePROfiHdrk4PHUJQ0RtYQ1xEZ-A1-z25gqRsVHX7np-AH8VZqY |
| linkProvider | Cornell University |
| openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Inexact+SARAH+Algorithm+for+Stochastic+Optimization&rft.au=Nguyen%2C+Lam+M&rft.au=Scheinberg%2C+Katya&rft.au=Tak%C3%A1%C4%8D%2C+Martin&rft.date=2018-11-25&rft_id=info:doi/10.48550%2Farxiv.1811.10105&rft.externalDocID=1811_10105 |