Inexact SARAH Algorithm for Stochastic Optimization

We develop and analyze a variant of the SARAH algorithm, which does not require computation of the exact gradient. Thus this new method can be applied to general expectation minimization problems rather than only finite sum problems. While the original SARAH algorithm, as well as its predecessor, SV...

Full description

Saved in:
Bibliographic Details
Main Authors Nguyen, Lam M, Scheinberg, Katya, Takáč, Martin
Format Journal Article
LanguageEnglish
Published 25.11.2018
Subjects
Online AccessGet full text
DOI10.48550/arxiv.1811.10105

Cover

Abstract We develop and analyze a variant of the SARAH algorithm, which does not require computation of the exact gradient. Thus this new method can be applied to general expectation minimization problems rather than only finite sum problems. While the original SARAH algorithm, as well as its predecessor, SVRG, require an exact gradient computation on each outer iteration, the inexact variant of SARAH (iSARAH), which we develop here, requires only stochastic gradient computed on a mini-batch of sufficient size. The proposed method combines variance reduction via sample size selection and iterative stochastic gradient updates. We analyze the convergence rate of the algorithms for strongly convex and non-strongly convex cases, under smooth assumption with appropriate mini-batch size selected for each case. We show that with an additional, reasonable, assumption iSARAH achieves the best known complexity among stochastic methods in the case of non-strongly convex stochastic functions.
AbstractList We develop and analyze a variant of the SARAH algorithm, which does not require computation of the exact gradient. Thus this new method can be applied to general expectation minimization problems rather than only finite sum problems. While the original SARAH algorithm, as well as its predecessor, SVRG, require an exact gradient computation on each outer iteration, the inexact variant of SARAH (iSARAH), which we develop here, requires only stochastic gradient computed on a mini-batch of sufficient size. The proposed method combines variance reduction via sample size selection and iterative stochastic gradient updates. We analyze the convergence rate of the algorithms for strongly convex and non-strongly convex cases, under smooth assumption with appropriate mini-batch size selected for each case. We show that with an additional, reasonable, assumption iSARAH achieves the best known complexity among stochastic methods in the case of non-strongly convex stochastic functions.
Author Nguyen, Lam M
Scheinberg, Katya
Takáč, Martin
Author_xml – sequence: 1
  givenname: Lam M
  surname: Nguyen
  fullname: Nguyen, Lam M
– sequence: 2
  givenname: Katya
  surname: Scheinberg
  fullname: Scheinberg, Katya
– sequence: 3
  givenname: Martin
  surname: Takáč
  fullname: Takáč, Martin
BackLink https://doi.org/10.48550/arXiv.1811.10105$$DView paper in arXiv
BookMark eNrjYmDJy89LZWCQNDTQM7EwNTXQTyyqyCzTM7QwNNQzNDA0MOVkMPbMS61ITC5RCHYMcvRQcMxJzy_KLMnIVUjLL1IILslPzkgsLslMVvAvKMnMzaxKLMnMz-NhYE1LzClO5YXS3Azybq4hzh66YOPjC4oycxOLKuNB1sSDrTEmrAIA4ywzQA
ContentType Journal Article
Copyright http://arxiv.org/licenses/nonexclusive-distrib/1.0
Copyright_xml – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0
DBID AKY
AKZ
GOX
DOI 10.48550/arxiv.1811.10105
DatabaseName arXiv Computer Science
arXiv Mathematics
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 1811_10105
GroupedDBID AKY
AKZ
GOX
ID FETCH-arxiv_primary_1811_101053
IEDL.DBID GOX
IngestDate Tue Jul 22 23:02:37 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-arxiv_primary_1811_101053
OpenAccessLink https://arxiv.org/abs/1811.10105
ParticipantIDs arxiv_primary_1811_10105
PublicationCentury 2000
PublicationDate 2018-11-25
PublicationDateYYYYMMDD 2018-11-25
PublicationDate_xml – month: 11
  year: 2018
  text: 2018-11-25
  day: 25
PublicationDecade 2010
PublicationYear 2018
Score 3.350194
SecondaryResourceType preprint
Snippet We develop and analyze a variant of the SARAH algorithm, which does not require computation of the exact gradient. Thus this new method can be applied to...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Learning
Mathematics - Optimization and Control
Title Inexact SARAH Algorithm for Stochastic Optimization
URI https://arxiv.org/abs/1811.10105
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdVxBS8MwGP2YPe0iyhxVp-bgtYw2ycyOZViqoAWn0FtJ0nQV3Cazk_58v6QVveyahOQj4eO9R3gP4BYRSynEiaDkhgeMSoo9V7JASC1oNeORptY7_PQ8S9_YY87zAZBfL4zcte_fXT6w-poi_IRWX9qQ0iMkCtbMm-Xd56SL4urX_61DjumG_oFEcgLHPbsjcfccpzAwmxHQh41ppW7IMn6JUxJ_rLaoyOs1Qb5Ils1W19KGJZMMu3fd2yLP4Ca5f12kgTum-OwyIQpbQeEqoGPwULkbHwhDuXAnDBc61GweaVWGjFeKVnJONePROfiHdrk4PHUJQ0RtYQ1xEZ-A1-z25gqRsVHX7np-AH8VZqY
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Inexact+SARAH+Algorithm+for+Stochastic+Optimization&rft.au=Nguyen%2C+Lam+M&rft.au=Scheinberg%2C+Katya&rft.au=Tak%C3%A1%C4%8D%2C+Martin&rft.date=2018-11-25&rft_id=info:doi/10.48550%2Farxiv.1811.10105&rft.externalDocID=1811_10105