Double Sparse Deep Reinforcement Learning via Multilayer Sparse Coding and Nonconvex Regularized Pruning
Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters,...
Saved in:
| Published in | IEEE transactions on cybernetics Vol. 53; no. 2; pp. 765 - 778 |
|---|---|
| Main Authors | , , , , |
| Format | Journal Article |
| Language | English |
| Published |
United States
IEEE
01.02.2023
The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
| Subjects | |
| Online Access | Get full text |
| ISSN | 2168-2267 2168-2275 2168-2275 |
| DOI | 10.1109/TCYB.2022.3157892 |
Cover
| Abstract | Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep <inline-formula> <tex-math notation="LaTeX">q </tex-math></inline-formula> network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations. |
|---|---|
| AbstractList | Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep <inline-formula> <tex-math notation="LaTeX">q </tex-math></inline-formula> network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations. Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep q network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations. Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep [Formula Omitted] network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations. Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep q network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations.Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems. However, the process of acquiring representations in DRL is easily affected by interference from models, and moreover leaves unnecessary parameters, leading to control performance reduction. In this article, we propose a double sparse DRL via multilayer sparse coding and nonconvex regularized pruning. To alleviate interference in DRL, we propose a multilayer sparse-coding-structural network to obtain deep sparse representation for control in reinforcement learning. Furthermore, we employ a nonconvex log regularizer to promote strong sparsity, efficiently removing the unnecessary weights with a regularizer-based pruning scheme. Hence, a double sparse DRL algorithm is developed, which can not only learn deep sparse representation to reduce the interference but also remove redundant weights while keeping the robust performance. The experimental results in five benchmark environments of the deep q network (DQN) architecture demonstrate that the proposed method with deep sparse representations from the multilayer sparse-coding structure can outperform existing sparse-coding-based DRL in control, for example, completing Mountain Car with 140.81 steps, achieving near 10% reward increase from the single-layer sparse-coding DRL algorithm, and obtaining 286.08 scores in Catcher, which are over two times the rewards of the other algorithms. Moreover, the proposed algorithm can reduce over 80% parameters while keeping performance improvements from deep sparse representations. |
| Author | Chen, Wuhui Wu, Jiqiang Li, Zhenni Zhao, Haoli Zheng, Zibin |
| Author_xml | – sequence: 1 givenname: Haoli orcidid: 0000-0002-4004-509X surname: Zhao fullname: Zhao, Haoli email: zhaohli1989@hotmail.com organization: School of Automation, Guangdong University of Technology, Guangzhou, China – sequence: 2 givenname: Jiqiang surname: Wu fullname: Wu, Jiqiang email: wujq27@mail2.sysu.edu.cn organization: School of Automation, Guangdong University of Technology, Guangzhou, China – sequence: 3 givenname: Zhenni orcidid: 0000-0001-8098-0341 surname: Li fullname: Li, Zhenni email: lizhenni2012@gmail.com organization: School of Automation and the Guangdong Key Laboratory of IoT Information Technology, Guangdong University of Technology, Guangzhou, China – sequence: 4 givenname: Wuhui orcidid: 0000-0003-4430-7904 surname: Chen fullname: Chen, Wuhui email: chenwuh@mail.sysu.edu.cn organization: School of Automation, Guangdong University of Technology, Guangzhou, China – sequence: 5 givenname: Zibin orcidid: 0000-0001-7872-7718 surname: Zheng fullname: Zheng, Zibin email: zhzibin@mail.sysu.edu.cn organization: School of Automation, Guangdong University of Technology, Guangzhou, China |
| BackLink | https://www.ncbi.nlm.nih.gov/pubmed/35316206$$D View this record in MEDLINE/PubMed |
| BookMark | eNp9kUtv1DAURi1UREvpD0BIKBIbNjPY169kCVMKSMNDUBasLMe5Ka4y9mAnVcuvx9FMZ9EF3tiyz7m-9veUHIUYkJDnjC4Zo82by9Wvd0ugAEvOpK4beEROgKl6AaDl0WGt9DE5y_mallGXraZ-Qo655EwBVSfk93mc2gGrH1ubMlbniNvqO_rQx-Rwg2Gs1mhT8OGquvG2-jwNox_sHaZ7YxW7-dCGrvoSg4vhBm9LhatpsMn_xa76lqZZf0Ye93bIeLafT8nPi_eXq4-L9dcPn1Zv1wsnmBwXdd8jE045sE3fQwe2dN1yycpDW9rVQkjXgdTA0QKHwkglaNuD7bhmgvNT8npXd5vinwnzaDY-OxwGGzBO2YASwDlVUBf01QP0Ok4plO4MaCWZlrShhXq5p6Z2g53ZJr-x6c7c_2EB2A5wKeacsD8gjJo5KjNHZeaozD6q4ugHjvOjHX0MY7J--K_5Ymd6RDzc1GhBqaD8H1fXn20 |
| CODEN | ITCEB8 |
| CitedBy_id | crossref_primary_10_1007_s11042_024_19192_x crossref_primary_10_1016_j_automatica_2024_111762 crossref_primary_10_33012_navi_667 crossref_primary_10_1016_j_rineng_2024_101751 crossref_primary_10_1109_TVT_2024_3399826 crossref_primary_10_26599_TST_2024_9010004 crossref_primary_10_3390_app13053125 crossref_primary_10_1016_j_neucom_2023_126789 crossref_primary_10_1109_JIOT_2023_3345943 crossref_primary_10_1002_ett_4929 crossref_primary_10_1016_j_neucom_2024_129139 crossref_primary_10_1109_TCYB_2024_3385910 crossref_primary_10_1109_JSAC_2022_3213283 crossref_primary_10_3390_electronics13214168 crossref_primary_10_1109_TAES_2023_3342794 crossref_primary_10_14326_abe_13_134 crossref_primary_10_1007_s10489_024_05464_4 crossref_primary_10_1109_TMC_2023_3320106 crossref_primary_10_1063_5_0255692 |
| Cites_doi | 10.24963/ijcai.2020/370 10.1109/TCYB.2019.2890974 10.1109/TCYB.2019.2950105 10.24963/ijcai.2017/287 10.1007/s10462-011-9289-8 10.1109/JIOT.2020.3016644 10.1007/s11548-019-02030-z 10.24963/ijcai.2020/396 10.1609/aaai.v34i04.5766 10.1007/BF00115009 10.1109/JSTSP.2020.2967566 10.1109/ICASSP.2014.6854997 10.1109/TVT.2019.2924015 10.1109/JSEN.2017.2736641 10.24963/ijcai.2019/379 10.1007/BF00992698 10.1109/TPAMI.2019.2904255 10.1109/TSP.2018.2846226 10.1038/nature14236 10.1609/aaai.v33i01.33014384 10.1117/1.JMI.2.2.024006 10.1609/aaai.v34i04.5963 |
| ContentType | Journal Article |
| Copyright | Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023 |
| Copyright_xml | – notice: Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023 |
| DBID | 97E RIA RIE AAYXX CITATION NPM 7SC 7SP 7TB 8FD F28 FR3 H8D JQ2 L7M L~C L~D 7X8 |
| DOI | 10.1109/TCYB.2022.3157892 |
| DatabaseName | IEEE All-Society Periodicals Package (ASPP) 2005–Present IEEE All-Society Periodicals Package (ASPP) 1998–Present IEEE Electronic Library (IEL) CrossRef PubMed Computer and Information Systems Abstracts Electronics & Communications Abstracts Mechanical & Transportation Engineering Abstracts Technology Research Database ANTE: Abstracts in New Technology & Engineering Engineering Research Database Aerospace Database ProQuest Computer Science Collection Advanced Technologies Database with Aerospace Computer and Information Systems Abstracts Academic Computer and Information Systems Abstracts Professional MEDLINE - Academic |
| DatabaseTitle | CrossRef PubMed Aerospace Database Technology Research Database Computer and Information Systems Abstracts – Academic Mechanical & Transportation Engineering Abstracts Electronics & Communications Abstracts ProQuest Computer Science Collection Computer and Information Systems Abstracts Engineering Research Database Advanced Technologies Database with Aerospace ANTE: Abstracts in New Technology & Engineering Computer and Information Systems Abstracts Professional MEDLINE - Academic |
| DatabaseTitleList | PubMed Aerospace Database MEDLINE - Academic |
| Database_xml | – sequence: 1 dbid: NPM name: PubMed url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed sourceTypes: Index Database – sequence: 2 dbid: RIE name: IEEE Xplore (NTUSG) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
| DeliveryMethod | fulltext_linktorsrc |
| Discipline | Sciences (General) |
| EISSN | 2168-2275 |
| EndPage | 778 |
| ExternalDocumentID | 35316206 10_1109_TCYB_2022_3157892 9740040 |
| Genre | orig-research Journal Article |
| GrantInformation_xml | – fundername: Guangzhou Science and Technology Program Project grantid: 202002030289; 6142006200403 – fundername: National Key Research and Development Plan grantid: 2021YFB2700302 – fundername: National Natural Science Foundation of China grantid: 62172453; 61727810; 61803096; 62073086; 62076077; U191140003 funderid: 10.13039/501100001809 – fundername: Program for Guangdong Introducing Innovative and Entrepreneurial Teams grantid: 2017ZT07X355 – fundername: Guangdong Provincial Pearl River Talents Program; Pearl River Talent Recruitment Program grantid: 2019QN01X130 funderid: 10.13039/100016691 |
| GroupedDBID | 0R~ 4.4 6IK 97E AAJGR AARMG AASAJ AAWTH ABAZT ABQJQ ABVLG ACIWK AENEX AGQYO AGSQL AHBIQ AKJIK AKQYR ALMA_UNASSIGNED_HOLDINGS ATWAV BEFXN BFFAM BGNUA BKEBE BPEOZ EBS EJD HZ~ IFIPE IPLJI JAVBF M43 O9- OCL PQQKQ RIA RIE RNS AAYXX CITATION NPM RIG 7SC 7SP 7TB 8FD F28 FR3 H8D JQ2 L7M L~C L~D 7X8 |
| ID | FETCH-LOGICAL-c415t-8ffe14c6c2a9ff2d2a081b351022b0d8445cd25723ea2329ff5640bf2ad371433 |
| IEDL.DBID | RIE |
| ISSN | 2168-2267 2168-2275 |
| IngestDate | Sun Sep 28 11:47:34 EDT 2025 Sun Oct 05 00:03:16 EDT 2025 Thu Apr 03 07:03:20 EDT 2025 Wed Oct 01 01:36:44 EDT 2025 Thu Apr 24 23:12:04 EDT 2025 Wed Aug 27 02:14:38 EDT 2025 |
| IsPeerReviewed | true |
| IsScholarly | true |
| Issue | 2 |
| Language | English |
| License | https://ieeexplore.ieee.org/Xplorehelp/downloads/license-information/IEEE.html https://doi.org/10.15223/policy-029 https://doi.org/10.15223/policy-037 |
| LinkModel | DirectLink |
| MergedId | FETCHMERGED-LOGICAL-c415t-8ffe14c6c2a9ff2d2a081b351022b0d8445cd25723ea2329ff5640bf2ad371433 |
| Notes | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 content type line 23 |
| ORCID | 0000-0001-7872-7718 0000-0002-4004-509X 0000-0001-8098-0341 0000-0003-4430-7904 |
| PMID | 35316206 |
| PQID | 2765175090 |
| PQPubID | 85422 |
| PageCount | 14 |
| ParticipantIDs | crossref_citationtrail_10_1109_TCYB_2022_3157892 ieee_primary_9740040 proquest_miscellaneous_2642330628 pubmed_primary_35316206 proquest_journals_2765175090 crossref_primary_10_1109_TCYB_2022_3157892 |
| ProviderPackageCode | CITATION AAYXX |
| PublicationCentury | 2000 |
| PublicationDate | 2023-02-01 |
| PublicationDateYYYYMMDD | 2023-02-01 |
| PublicationDate_xml | – month: 02 year: 2023 text: 2023-02-01 day: 01 |
| PublicationDecade | 2020 |
| PublicationPlace | United States |
| PublicationPlace_xml | – name: United States – name: Piscataway |
| PublicationTitle | IEEE transactions on cybernetics |
| PublicationTitleAbbrev | TCYB |
| PublicationTitleAlternate | IEEE Trans Cybern |
| PublicationYear | 2023 |
| Publisher | IEEE The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
| Publisher_xml | – name: IEEE – name: The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
| References | ref35 ref13 banerjee (ref34) 2005; 6 ref12 mnih (ref32) 2015; 518 ref31 ref30 ref33 ref10 ref2 ref17 ref18 tian (ref6) 2020; 2020 rusu (ref20) 2015 srivastava (ref36) 2014; 15 fan (ref28) 2020 zhou (ref1) 2020; 33 ref24 sutton (ref16) 2018 van hasselt (ref27) 2018 ref23 ref26 lv (ref3) 2020 hernandez-garcia (ref14) 2019 ref22 ref21 sutton (ref25) 2000 wang (ref11) 2020 ref29 ref8 ref7 li (ref15) 2021 ref9 ref4 yang (ref19) 2019 ref5 |
| References_xml | – ident: ref10 doi: 10.24963/ijcai.2020/370 – year: 2018 ident: ref27 article-title: Deep reinforcement learning and the deadly triad publication-title: arXiv 1812 02648 – start-page: 1 year: 2019 ident: ref19 article-title: DeepHoyer: Learning sparser neural network with differentiable scale-invariant sparsity measures publication-title: Proc Int Conf Learn Represent – ident: ref9 doi: 10.1109/TCYB.2019.2890974 – year: 2015 ident: ref20 article-title: Policy distillation publication-title: arXiv 1511 06295 – year: 2018 ident: ref16 publication-title: Reinforcement Learning An Introduction – ident: ref31 doi: 10.1109/TCYB.2019.2950105 – ident: ref12 doi: 10.24963/ijcai.2017/287 – ident: ref23 doi: 10.1007/s10462-011-9289-8 – year: 2021 ident: ref15 article-title: Accelerated log-regularized convolutional transform learning and its convergence guarantee publication-title: IEEE Trans Cybern – ident: ref2 doi: 10.1109/JIOT.2020.3016644 – start-page: 7968 year: 2020 ident: ref11 article-title: Improving generalization in reinforcement learning with mixture regularization publication-title: Proc Conf Neural Inf Process Syst – start-page: 1057 year: 2000 ident: ref25 article-title: Policy gradient methods for reinforcement learning with function approximation publication-title: Advances in neural information processing systems – ident: ref5 doi: 10.1007/s11548-019-02030-z – ident: ref8 doi: 10.24963/ijcai.2020/396 – ident: ref7 doi: 10.1609/aaai.v34i04.5766 – ident: ref33 doi: 10.1007/BF00115009 – ident: ref21 doi: 10.1109/JSTSP.2020.2967566 – ident: ref35 doi: 10.1109/ICASSP.2014.6854997 – ident: ref18 doi: 10.1109/TVT.2019.2924015 – ident: ref4 doi: 10.1109/JSEN.2017.2736641 – start-page: 486 year: 2020 ident: ref28 article-title: A theoretical analysis of deep Q-learning publication-title: Proc 2nd Conf Learn Dyn Control – ident: ref29 doi: 10.24963/ijcai.2019/379 – volume: 33 start-page: 13504 year: 2020 ident: ref1 article-title: Promoting stochasticity for expressive policies via a simple and efficient regularization method publication-title: Advances in neural information processing systems – ident: ref26 doi: 10.1007/BF00992698 – ident: ref30 doi: 10.1109/TPAMI.2019.2904255 – volume: 2020 start-page: 1 year: 2020 ident: ref6 article-title: Multi-step medical image segmentation based on reinforcement learning publication-title: J Ambient Intell Humanized Comput – ident: ref17 doi: 10.1109/TSP.2018.2846226 – volume: 15 start-page: 1929 year: 2014 ident: ref36 article-title: Dropout: A simple way to prevent neural networks from overfitting publication-title: J Mach Learn Res – volume: 518 start-page: 529 year: 2015 ident: ref32 article-title: Human-level control through deep reinforcement learning publication-title: Nature doi: 10.1038/nature14236 – volume: 6 start-page: 1705 year: 2005 ident: ref34 article-title: Clustering with Bregman divergences publication-title: J Mach Learn Res – year: 2020 ident: ref3 article-title: Integrated double estimator architecture for reinforcement learning publication-title: IEEE Trans Cybern – ident: ref13 doi: 10.1609/aaai.v33i01.33014384 – year: 2019 ident: ref14 article-title: Learning sparse representations incrementally in deep reinforcement learning publication-title: arXiv 1912 04002 – ident: ref24 doi: 10.1117/1.JMI.2.2.024006 – ident: ref22 doi: 10.1609/aaai.v34i04.5963 |
| SSID | ssj0000816898 |
| Score | 2.4800391 |
| Snippet | Deep reinforcement learning (DRL), which highly depends on the data representation, has shown its potential in many practical decision-making problems.... |
| SourceID | proquest pubmed crossref ieee |
| SourceType | Aggregation Database Index Database Enrichment Source Publisher |
| StartPage | 765 |
| SubjectTerms | Algorithms Coding Decision making Deep learning Deep reinforcement learning (DRL) Encoding Feature extraction Image segmentation Interference log regularizer Machine learning Monolayers multilayer sparse coding Multilayers Nonhomogeneous media Parameters pruning Reinforcement learning Representations Training |
| Title | Double Sparse Deep Reinforcement Learning via Multilayer Sparse Coding and Nonconvex Regularized Pruning |
| URI | https://ieeexplore.ieee.org/document/9740040 https://www.ncbi.nlm.nih.gov/pubmed/35316206 https://www.proquest.com/docview/2765175090 https://www.proquest.com/docview/2642330628 |
| Volume | 53 |
| hasFullText | 1 |
| inHoldings | 1 |
| isFullTextHit | |
| isPrint | |
| journalDatabaseRights | – providerCode: PRVIEE databaseName: IEEE Xplore (NTUSG) customDbUrl: eissn: 2168-2275 dateEnd: 99991231 omitProxy: false ssIdentifier: ssj0000816898 issn: 2168-2267 databaseCode: RIE dateStart: 20130101 isFulltext: true titleUrlDefault: https://ieeexplore.ieee.org/ providerName: IEEE |
| link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3dT9RAEJ8AT74oiB8nYNbEBzX26O222_ZRTwkxgRiFBJ-a3e0UiJf2AldC-OuZ2W4bY9T41mQ_-jGznd_szswP4HWRudylhY20RYwSrePIZNJFhpSH0Hduc1-34OhYH54mX87SszV4P-bCIKIPPsMpX_qz_Kp1HW-V7RP2ZaVbh_Us132u1rif4gkkPPWtpIuIUEUWDjFncbF_Mv_xkZxBKclHJR0tmMRGkfppyVRHv1gkT7Hyd7Tprc7BIzganrcPNvk57VZ26u5-K-X4vy-0CQ8D_BQfen3ZgjVsHsNWWODX4k2oQv12Gy4IWtsFiu9Lcn1RfEJcim_o66w6v6UoQmnWc3FzaYRP5F0YAvDDiHnLZlGYphLHbeOj229phnOOe728w0p8vep4-BM4Pfh8Mj-MAi1D5Mjar6K8rnGWOO2kKepaVtLQt7cqZd_RxlWeJKmr6E8gFRrCa9Qn1Ulsa2kqLg-o1FPYaNoGn4NALTlsJZ85myVYJ-QMpoqmLuK0zoxVE4gH0ZQu1Cxn6oxF6X2XuChZsCULtgyCncC7cciyL9jxr87bLJSxY5DHBHYH-ZdhSV-XMtPpjPEVNb8am2kx8gmLabDtqA95c0pxWuoEnvV6M849qNuLP99zBx4wk30fEL4LG6urDvcI76zsS6_o94Cd-Aw |
| linkProvider | IEEE |
| linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1Lb9QwEB6VcoALUMpjoYCROAAi26wT53GEhWqB7grBVmpPke1MSsUqWbUbhPrrmXGcCCFA3CL5kceMM9_YM_MBPMtTm1mVmyAxiEGcJGGgU2kDTcpD6DszmatbMF8ks6P4w7E63oJXQy4MIrrgMxzzpTvLLxvb8lbZPmFfVrorcFXFcay6bK1hR8VRSDjyW0kXAeGK1B9jTsJ8fzk9eUPuoJTkpZKW5kxjE5ECJpLJjn6xSY5k5e9409mdg5sw75-4Czf5Nm43Zmwvfyvm-L-vdAtueAAqXncaswNbWN-GHb_EL8RzX4f6xS58JXBtVii-rMn5RfEWcS0-o6u0at2movDFWU_F9zMtXCrvShOE70dMGzaMQtelWDS1i2__QTOccuTr2SWW4tN5y8PvwNHBu-V0FnhihsCSvd8EWVXhJLaJlTqvKllKTd_eRIq9RxOWGcnGlvQvkBFqQmzURyVxaCqpSy4QGEV3YbtuarwPAhPJgSvZxJo0xiomd1BFNHUeqirVJhpB2IumsL5qOZNnrArnvYR5wYItWLCFF-wIXg5D1l3Jjn913mWhDB29PEaw18u_8Iv6opBpoiaMsKj56dBMy5HPWHSNTUt9yJ-LIk5MHcG9Tm-GuXt1e_Dnez6Ba7Pl_LA4fL_4-BCuM699Fx6-B9ub8xYfEfrZmMdO6X8Coyj7WQ |
| openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Double+Sparse+Deep+Reinforcement+Learning+via+Multilayer+Sparse+Coding+and+Nonconvex+Regularized+Pruning&rft.jtitle=IEEE+transactions+on+cybernetics&rft.au=Zhao%2C+Haoli&rft.au=Wu%2C+Jiqiang&rft.au=Li%2C+Zhenni&rft.au=Chen%2C+Wuhui&rft.date=2023-02-01&rft.issn=2168-2275&rft.eissn=2168-2275&rft.volume=53&rft.issue=2&rft.spage=765&rft_id=info:doi/10.1109%2FTCYB.2022.3157892&rft.externalDBID=NO_FULL_TEXT |
| thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2168-2267&client=summon |
| thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2168-2267&client=summon |
| thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2168-2267&client=summon |