Double Sparse Deep Reinforcement Learning via Multilayer Sparse Coding and Nonconvex Regularized Pruning

Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters,...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on cybernetics Vol. 53; no. 2; pp. 765 - 778
Main Authors Zhao, Haoli, Wu, Jiqiang, Li, Zhenni, Chen, Wuhui, Zheng, Zibin
Format Journal Article
LanguageEnglish
Published United States IEEE 01.02.2023
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text
ISSN2168-2267
2168-2275
2168-2275
DOI10.1109/TCYB.2022.3157892

Cover

Abstract Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep <inline-formula> <tex-math notation="LaTeX">q </tex-math></inline-formula> network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations.
AbstractList Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep <inline-formula> <tex-math notation="LaTeX">q </tex-math></inline-formula> network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations.
Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep q network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations.
Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep [Formula Omitted] network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations.
Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep q network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations.Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep q network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations.
Author Chen, Wuhui
Wu, Jiqiang
Li, Zhenni
Zhao, Haoli
Zheng, Zibin
Author_xml – sequence: 1
  givenname: Haoli
  orcidid: 0000-0002-4004-509X
  surname: Zhao
  fullname: Zhao, Haoli
  email: zhaohli1989@hotmail.com
  organization: School of Automation, Guangdong University of Technology, Guangzhou, China
– sequence: 2
  givenname: Jiqiang
  surname: Wu
  fullname: Wu, Jiqiang
  email: wujq27@mail2.sysu.edu.cn
  organization: School of Automation, Guangdong University of Technology, Guangzhou, China
– sequence: 3
  givenname: Zhenni
  orcidid: 0000-0001-8098-0341
  surname: Li
  fullname: Li, Zhenni
  email: lizhenni2012@gmail.com
  organization: School of Automation and the Guangdong Key Laboratory of IoT Information Technology, Guangdong University of Technology, Guangzhou, China
– sequence: 4
  givenname: Wuhui
  orcidid: 0000-0003-4430-7904
  surname: Chen
  fullname: Chen, Wuhui
  email: chenwuh@mail.sysu.edu.cn
  organization: School of Automation, Guangdong University of Technology, Guangzhou, China
– sequence: 5
  givenname: Zibin
  orcidid: 0000-0001-7872-7718
  surname: Zheng
  fullname: Zheng, Zibin
  email: zhzibin@mail.sysu.edu.cn
  organization: School of Automation, Guangdong University of Technology, Guangzhou, China
BackLink https://www.ncbi.nlm.nih.gov/pubmed/35316206$$D View this record in MEDLINE/PubMed
BookMark eNp9kUtv1DAURi1UREvpD0BIKBIbNjPY169kCVMKSMNDUBasLMe5Ka4y9mAnVcuvx9FMZ9EF3tiyz7m-9veUHIUYkJDnjC4Zo82by9Wvd0ugAEvOpK4beEROgKl6AaDl0WGt9DE5y_mallGXraZ-Qo655EwBVSfk93mc2gGrH1ubMlbniNvqO_rQx-Rwg2Gs1mhT8OGquvG2-jwNox_sHaZ7YxW7-dCGrvoSg4vhBm9LhatpsMn_xa76lqZZf0Ye93bIeLafT8nPi_eXq4-L9dcPn1Zv1wsnmBwXdd8jE045sE3fQwe2dN1yycpDW9rVQkjXgdTA0QKHwkglaNuD7bhmgvNT8npXd5vinwnzaDY-OxwGGzBO2YASwDlVUBf01QP0Ok4plO4MaCWZlrShhXq5p6Z2g53ZJr-x6c7c_2EB2A5wKeacsD8gjJo5KjNHZeaozD6q4ugHjvOjHX0MY7J--K_5Ymd6RDzc1GhBqaD8H1fXn20
CODEN ITCEB8
CitedBy_id crossref_primary_10_1007_s11042_024_19192_x
crossref_primary_10_1016_j_automatica_2024_111762
crossref_primary_10_33012_navi_667
crossref_primary_10_1016_j_rineng_2024_101751
crossref_primary_10_1109_TVT_2024_3399826
crossref_primary_10_26599_TST_2024_9010004
crossref_primary_10_3390_app13053125
crossref_primary_10_1016_j_neucom_2023_126789
crossref_primary_10_1109_JIOT_2023_3345943
crossref_primary_10_1002_ett_4929
crossref_primary_10_1016_j_neucom_2024_129139
crossref_primary_10_1109_TCYB_2024_3385910
crossref_primary_10_1109_JSAC_2022_3213283
crossref_primary_10_3390_electronics13214168
crossref_primary_10_1109_TAES_2023_3342794
crossref_primary_10_14326_abe_13_134
crossref_primary_10_1007_s10489_024_05464_4
crossref_primary_10_1109_TMC_2023_3320106
crossref_primary_10_1063_5_0255692
Cites_doi 10.24963/ijcai.2020/370
10.1109/TCYB.2019.2890974
10.1109/TCYB.2019.2950105
10.24963/ijcai.2017/287
10.1007/s10462-011-9289-8
10.1109/JIOT.2020.3016644
10.1007/s11548-019-02030-z
10.24963/ijcai.2020/396
10.1609/aaai.v34i04.5766
10.1007/BF00115009
10.1109/JSTSP.2020.2967566
10.1109/ICASSP.2014.6854997
10.1109/TVT.2019.2924015
10.1109/JSEN.2017.2736641
10.24963/ijcai.2019/379
10.1007/BF00992698
10.1109/TPAMI.2019.2904255
10.1109/TSP.2018.2846226
10.1038/nature14236
10.1609/aaai.v33i01.33014384
10.1117/1.JMI.2.2.024006
10.1609/aaai.v34i04.5963
ContentType Journal Article
Copyright Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023
Copyright_xml – notice: Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023
DBID 97E
RIA
RIE
AAYXX
CITATION
NPM
7SC
7SP
7TB
8FD
F28
FR3
H8D
JQ2
L7M
L~C
L~D
7X8
DOI 10.1109/TCYB.2022.3157892
DatabaseName IEEE All-Society Periodicals Package (ASPP) 2005–Present
IEEE All-Society Periodicals Package (ASPP) 1998–Present
IEEE Electronic Library (IEL)
CrossRef
PubMed
Computer and Information Systems Abstracts
Electronics & Communications Abstracts
Mechanical & Transportation Engineering Abstracts
Technology Research Database
ANTE: Abstracts in New Technology & Engineering
Engineering Research Database
Aerospace Database
ProQuest Computer Science Collection
Advanced Technologies Database with Aerospace
Computer and Information Systems Abstracts – Academic
Computer and Information Systems Abstracts Professional
MEDLINE - Academic
DatabaseTitle CrossRef
PubMed
Aerospace Database
Technology Research Database
Computer and Information Systems Abstracts – Academic
Mechanical & Transportation Engineering Abstracts
Electronics & Communications Abstracts
ProQuest Computer Science Collection
Computer and Information Systems Abstracts
Engineering Research Database
Advanced Technologies Database with Aerospace
ANTE: Abstracts in New Technology & Engineering
Computer and Information Systems Abstracts Professional
MEDLINE - Academic
DatabaseTitleList
PubMed
Aerospace Database
MEDLINE - Academic
Database_xml – sequence: 1
  dbid: NPM
  name: PubMed
  url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
– sequence: 2
  dbid: RIE
  name: IEEE Xplore (NTUSG)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Sciences (General)
EISSN 2168-2275
EndPage 778
ExternalDocumentID 35316206
10_1109_TCYB_2022_3157892
9740040
Genre orig-research
Journal Article
GrantInformation_xml – fundername: Guangzhou Science and Technology Program Project
  grantid: 202002030289; 6142006200403
– fundername: National Key Research and Development Plan
  grantid: 2021YFB2700302
– fundername: National Natural Science Foundation of China
  grantid: 62172453; 61727810; 61803096; 62073086; 62076077; U191140003
  funderid: 10.13039/501100001809
– fundername: Program for Guangdong Introducing Innovative and Entrepreneurial Teams
  grantid: 2017ZT07X355
– fundername: Guangdong Provincial Pearl River Talents Program; Pearl River Talent Recruitment Program
  grantid: 2019QN01X130
  funderid: 10.13039/100016691
GroupedDBID 0R~
4.4
6IK
97E
AAJGR
AARMG
AASAJ
AAWTH
ABAZT
ABQJQ
ABVLG
ACIWK
AENEX
AGQYO
AGSQL
AHBIQ
AKJIK
AKQYR
ALMA_UNASSIGNED_HOLDINGS
ATWAV
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
EBS
EJD
HZ~
IFIPE
IPLJI
JAVBF
M43
O9-
OCL
PQQKQ
RIA
RIE
RNS
AAYXX
CITATION
NPM
RIG
7SC
7SP
7TB
8FD
F28
FR3
H8D
JQ2
L7M
L~C
L~D
7X8
ID FETCH-LOGICAL-c415t-8ffe14c6c2a9ff2d2a081b351022b0d8445cd25723ea2329ff5640bf2ad371433
IEDL.DBID RIE
ISSN 2168-2267
2168-2275
IngestDate Sun Sep 28 11:47:34 EDT 2025
Sun Oct 05 00:03:16 EDT 2025
Thu Apr 03 07:03:20 EDT 2025
Wed Oct 01 01:36:44 EDT 2025
Thu Apr 24 23:12:04 EDT 2025
Wed Aug 27 02:14:38 EDT 2025
IsPeerReviewed true
IsScholarly true
Issue 2
Language English
License https://ieeexplore.ieee.org/Xplorehelp/downloads/license-information/IEEE.html
https://doi.org/10.15223/policy-029
https://doi.org/10.15223/policy-037
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c415t-8ffe14c6c2a9ff2d2a081b351022b0d8445cd25723ea2329ff5640bf2ad371433
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ORCID 0000-0001-7872-7718
0000-0002-4004-509X
0000-0001-8098-0341
0000-0003-4430-7904
PMID 35316206
PQID 2765175090
PQPubID 85422
PageCount 14
ParticipantIDs crossref_citationtrail_10_1109_TCYB_2022_3157892
ieee_primary_9740040
proquest_miscellaneous_2642330628
pubmed_primary_35316206
proquest_journals_2765175090
crossref_primary_10_1109_TCYB_2022_3157892
ProviderPackageCode CITATION
AAYXX
PublicationCentury 2000
PublicationDate 2023-02-01
PublicationDateYYYYMMDD 2023-02-01
PublicationDate_xml – month: 02
  year: 2023
  text: 2023-02-01
  day: 01
PublicationDecade 2020
PublicationPlace United States
PublicationPlace_xml – name: United States
– name: Piscataway
PublicationTitle IEEE transactions on cybernetics
PublicationTitleAbbrev TCYB
PublicationTitleAlternate IEEE Trans Cybern
PublicationYear 2023
Publisher IEEE
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Publisher_xml – name: IEEE
– name: The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
References ref35
ref13
banerjee (ref34) 2005; 6
ref12
mnih (ref32) 2015; 518
ref31
ref30
ref33
ref10
ref2
ref17
ref18
tian (ref6) 2020; 2020
rusu (ref20) 2015
srivastava (ref36) 2014; 15
fan (ref28) 2020
zhou (ref1) 2020; 33
ref24
sutton (ref16) 2018
van hasselt (ref27) 2018
ref23
ref26
lv (ref3) 2020
hernandez-garcia (ref14) 2019
ref22
ref21
sutton (ref25) 2000
wang (ref11) 2020
ref29
ref8
ref7
li (ref15) 2021
ref9
ref4
yang (ref19) 2019
ref5
References_xml – ident: ref10
  doi: 10.24963/ijcai.2020/370
– year: 2018
  ident: ref27
  article-title: Deep reinforcement learning and the deadly triad
  publication-title: arXiv 1812 02648
– start-page: 1
  year: 2019
  ident: ref19
  article-title: DeepHoyer: Learning sparser neural network with differentiable scale-invariant sparsity measures
  publication-title: Proc Int Conf Learn Represent
– ident: ref9
  doi: 10.1109/TCYB.2019.2890974
– year: 2015
  ident: ref20
  article-title: Policy distillation
  publication-title: arXiv 1511 06295
– year: 2018
  ident: ref16
  publication-title: Reinforcement Learning An Introduction
– ident: ref31
  doi: 10.1109/TCYB.2019.2950105
– ident: ref12
  doi: 10.24963/ijcai.2017/287
– ident: ref23
  doi: 10.1007/s10462-011-9289-8
– year: 2021
  ident: ref15
  article-title: Accelerated log-regularized convolutional transform learning and its convergence guarantee
  publication-title: IEEE Trans Cybern
– ident: ref2
  doi: 10.1109/JIOT.2020.3016644
– start-page: 7968
  year: 2020
  ident: ref11
  article-title: Improving generalization in reinforcement learning with mixture regularization
  publication-title: Proc Conf Neural Inf Process Syst
– start-page: 1057
  year: 2000
  ident: ref25
  article-title: Policy gradient methods for reinforcement learning with function approximation
  publication-title: Advances in neural information processing systems
– ident: ref5
  doi: 10.1007/s11548-019-02030-z
– ident: ref8
  doi: 10.24963/ijcai.2020/396
– ident: ref7
  doi: 10.1609/aaai.v34i04.5766
– ident: ref33
  doi: 10.1007/BF00115009
– ident: ref21
  doi: 10.1109/JSTSP.2020.2967566
– ident: ref35
  doi: 10.1109/ICASSP.2014.6854997
– ident: ref18
  doi: 10.1109/TVT.2019.2924015
– ident: ref4
  doi: 10.1109/JSEN.2017.2736641
– start-page: 486
  year: 2020
  ident: ref28
  article-title: A theoretical analysis of deep Q-learning
  publication-title: Proc 2nd Conf Learn Dyn Control
– ident: ref29
  doi: 10.24963/ijcai.2019/379
– volume: 33
  start-page: 13504
  year: 2020
  ident: ref1
  article-title: Promoting stochasticity for expressive policies via a simple and efficient regularization method
  publication-title: Advances in neural information processing systems
– ident: ref26
  doi: 10.1007/BF00992698
– ident: ref30
  doi: 10.1109/TPAMI.2019.2904255
– volume: 2020
  start-page: 1
  year: 2020
  ident: ref6
  article-title: Multi-step medical image segmentation based on reinforcement learning
  publication-title: J Ambient Intell Humanized Comput
– ident: ref17
  doi: 10.1109/TSP.2018.2846226
– volume: 15
  start-page: 1929
  year: 2014
  ident: ref36
  article-title: Dropout: A simple way to prevent neural networks from overfitting
  publication-title: J Mach Learn Res
– volume: 518
  start-page: 529
  year: 2015
  ident: ref32
  article-title: Human-level control through deep reinforcement learning
  publication-title: Nature
  doi: 10.1038/nature14236
– volume: 6
  start-page: 1705
  year: 2005
  ident: ref34
  article-title: Clustering with Bregman divergences
  publication-title: J Mach Learn Res
– year: 2020
  ident: ref3
  article-title: Integrated double estimator architecture for reinforcement learning
  publication-title: IEEE Trans Cybern
– ident: ref13
  doi: 10.1609/aaai.v33i01.33014384
– year: 2019
  ident: ref14
  article-title: Learning sparse representations incrementally in deep reinforcement learning
  publication-title: arXiv 1912 04002
– ident: ref24
  doi: 10.1117/1.JMI.2.2.024006
– ident: ref22
  doi: 10.1609/aaai.v34i04.5963
SSID ssj0000816898
Score 2.4800391
Snippet Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems....
SourceID proquest
pubmed
crossref
ieee
SourceType Aggregation Database
Index Database
Enrichment Source
Publisher
StartPage 765
SubjectTerms Algorithms
Coding
Decision making
Deep learning
Deep reinforcement learning (DRL)
Encoding
Feature extraction
Image segmentation
Interference
log regularizer
Machine learning
Monolayers
multilayer sparse coding
Multilayers
Nonhomogeneous media
Parameters
pruning
Reinforcement learning
Representations
Training
Title Double Sparse Deep Reinforcement Learning via Multilayer Sparse Coding and Nonconvex Regularized Pruning
URI https://ieeexplore.ieee.org/document/9740040
https://www.ncbi.nlm.nih.gov/pubmed/35316206
https://www.proquest.com/docview/2765175090
https://www.proquest.com/docview/2642330628
Volume 53
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
journalDatabaseRights – providerCode: PRVIEE
  databaseName: IEEE Xplore (NTUSG)
  customDbUrl:
  eissn: 2168-2275
  dateEnd: 99991231
  omitProxy: false
  ssIdentifier: ssj0000816898
  issn: 2168-2267
  databaseCode: RIE
  dateStart: 20130101
  isFulltext: true
  titleUrlDefault: https://ieeexplore.ieee.org/
  providerName: IEEE
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3dT9RAEJ8AT74oiB8nYNbEBzX26O222_ZRTwkxgRiFBJ-a3e0UiJf2AldC-OuZ2W4bY9T41mQ_-jGznd_szswP4HWRudylhY20RYwSrePIZNJFhpSH0Hduc1-34OhYH54mX87SszV4P-bCIKIPPsMpX_qz_Kp1HW-V7RP2ZaVbh_Us132u1rif4gkkPPWtpIuIUEUWDjFncbF_Mv_xkZxBKclHJR0tmMRGkfppyVRHv1gkT7Hyd7Tprc7BIzganrcPNvk57VZ26u5-K-X4vy-0CQ8D_BQfen3ZgjVsHsNWWODX4k2oQv12Gy4IWtsFiu9Lcn1RfEJcim_o66w6v6UoQmnWc3FzaYRP5F0YAvDDiHnLZlGYphLHbeOj229phnOOe728w0p8vep4-BM4Pfh8Mj-MAi1D5Mjar6K8rnGWOO2kKepaVtLQt7cqZd_RxlWeJKmr6E8gFRrCa9Qn1Ulsa2kqLg-o1FPYaNoGn4NALTlsJZ85myVYJ-QMpoqmLuK0zoxVE4gH0ZQu1Cxn6oxF6X2XuChZsCULtgyCncC7cciyL9jxr87bLJSxY5DHBHYH-ZdhSV-XMtPpjPEVNb8am2kx8gmLabDtqA95c0pxWuoEnvV6M849qNuLP99zBx4wk30fEL4LG6urDvcI76zsS6_o94Cd-Aw
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1Lb9QwEB6VcoALUMpjoYCROAAi26wT53GEhWqB7grBVmpPke1MSsUqWbUbhPrrmXGcCCFA3CL5kceMM9_YM_MBPMtTm1mVmyAxiEGcJGGgU2kDTcpD6DszmatbMF8ks6P4w7E63oJXQy4MIrrgMxzzpTvLLxvb8lbZPmFfVrorcFXFcay6bK1hR8VRSDjyW0kXAeGK1B9jTsJ8fzk9eUPuoJTkpZKW5kxjE5ECJpLJjn6xSY5k5e9409mdg5sw75-4Czf5Nm43Zmwvfyvm-L-vdAtueAAqXncaswNbWN-GHb_EL8RzX4f6xS58JXBtVii-rMn5RfEWcS0-o6u0at2movDFWU_F9zMtXCrvShOE70dMGzaMQtelWDS1i2__QTOccuTr2SWW4tN5y8PvwNHBu-V0FnhihsCSvd8EWVXhJLaJlTqvKllKTd_eRIq9RxOWGcnGlvQvkBFqQmzURyVxaCqpSy4QGEV3YbtuarwPAhPJgSvZxJo0xiomd1BFNHUeqirVJhpB2IumsL5qOZNnrArnvYR5wYItWLCFF-wIXg5D1l3Jjn913mWhDB29PEaw18u_8Iv6opBpoiaMsKj56dBMy5HPWHSNTUt9yJ-LIk5MHcG9Tm-GuXt1e_Dnez6Ba7Pl_LA4fL_4-BCuM699Fx6-B9ub8xYfEfrZmMdO6X8Coyj7WQ
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Double+Sparse+Deep+Reinforcement+Learning+via+Multilayer+Sparse+Coding+and+Nonconvex+Regularized+Pruning&rft.jtitle=IEEE+transactions+on+cybernetics&rft.au=Zhao%2C+Haoli&rft.au=Wu%2C+Jiqiang&rft.au=Li%2C+Zhenni&rft.au=Chen%2C+Wuhui&rft.date=2023-02-01&rft.issn=2168-2275&rft.eissn=2168-2275&rft.volume=53&rft.issue=2&rft.spage=765&rft_id=info:doi/10.1109%2FTCYB.2022.3157892&rft.externalDBID=NO_FULL_TEXT
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2168-2267&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2168-2267&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2168-2267&client=summon