Deep Learning in Reconfigurable Hardware: A Survey
Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related to the size and complexity of the networks and training algorithms. However, these structures are considerably resource-consuming and demand...
Saved in:
Published in | 2019 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) pp. 95 - 98 |
---|---|
Main Authors | , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
01.05.2019
|
Subjects | |
Online Access | Get full text |
DOI | 10.1109/IPDPSW.2019.00026 |
Cover
Abstract | Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related to the size and complexity of the networks and training algorithms. However, these structures are considerably resource-consuming and demand extra effort to be used on embedded systems. Researchers have chosen different alternatives to execute deep learning algorithms as servers, clusters, GPUs and FPGAs. Specific hardware structures have been designed to solve these problems in different platforms such as FPGAs and even ASICs. Although there are surveys about this subject, they do not present a clear criteria to real implementations, detailed explanation of design techniques and applied metrics. Presenting a different approach, this work analyzes reconfigurable hardware based structures designed to optimize deep learning algorithms. This work includes results of hardware accelerators implemented and tested using commercially available development boards avoiding simulation-based results.Results of the proposed analysis showed that a lot of effort is directed to this subject but presented results are far from what is expected for the performance of a hardware structure for deep learning. |
---|---|
AbstractList | Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related to the size and complexity of the networks and training algorithms. However, these structures are considerably resource-consuming and demand extra effort to be used on embedded systems. Researchers have chosen different alternatives to execute deep learning algorithms as servers, clusters, GPUs and FPGAs. Specific hardware structures have been designed to solve these problems in different platforms such as FPGAs and even ASICs. Although there are surveys about this subject, they do not present a clear criteria to real implementations, detailed explanation of design techniques and applied metrics. Presenting a different approach, this work analyzes reconfigurable hardware based structures designed to optimize deep learning algorithms. This work includes results of hardware accelerators implemented and tested using commercially available development boards avoiding simulation-based results.Results of the proposed analysis showed that a lot of effort is directed to this subject but presented results are far from what is expected for the performance of a hardware structure for deep learning. |
Author | Dias, Mauricio A. Ferreira, Daniel A.P. |
Author_xml | – sequence: 1 givenname: Mauricio A. surname: Dias fullname: Dias, Mauricio A. email: macdias@fho.edu.br organization: University Center of Heminio Ometto Foundation – sequence: 2 givenname: Daniel A.P. surname: Ferreira fullname: Ferreira, Daniel A.P. email: danielferreira@fho.edu.br organization: University Center of Heminio Ometto Foundation |
BookMark | eNotjMFKAzEUACMoaGs_QLzkB7a-l7e72XgrrdrCgsUqHkuyeSmRmpasVfr3inoamIEZiNO0SyzEFcIYEczNYjlbrl7HCtCMAUDVJ2KAWjVI1U8-F6O-f_v1TQUGLoSaMe9lyzanmDYyJvnE3S6FuDlk67Ys5zb7L5v5Vk7k6pA_-XgpzoLd9jz651C83N89T-dF-_iwmE7aIqKuPgpfedJEwFS7skMf6hJKdAF9yS4Yj51p2GmmACV3xihHGmp2Hi0jOKShuP77RmZe73N8t_m4brRuiBR9A-pxRE4 |
CODEN | IEEPAD |
ContentType | Conference Proceeding |
DBID | 6IE 6IL CBEJK RIE RIL |
DOI | 10.1109/IPDPSW.2019.00026 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Xplore POP ALL IEEE Xplore All Conference Proceedings IEEE Electronic Library (IEL) IEEE Proceedings Order Plans (POP All) 1998-Present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library (IEL) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
EISBN | 1728135109 9781728135106 |
EndPage | 98 |
ExternalDocumentID | 8778332 |
Genre | orig-research |
GroupedDBID | 6IE 6IF 6IL 6IN AAJGR AAWTH ABLEC ALMA_UNASSIGNED_HOLDINGS BEFXN BFFAM BGNUA BKEBE BPEOZ CBEJK IEGSK OCL RIB RIC RIE RIL |
ID | FETCH-LOGICAL-i175t-d5d37330e36b4c1df64041bf1d4ebf9d1c98eb7e3f04ec992b3706ebd1ae10b13 |
IEDL.DBID | RIE |
IngestDate | Wed Aug 27 02:40:28 EDT 2025 |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i175t-d5d37330e36b4c1df64041bf1d4ebf9d1c98eb7e3f04ec992b3706ebd1ae10b13 |
PageCount | 4 |
ParticipantIDs | ieee_primary_8778332 |
PublicationCentury | 2000 |
PublicationDate | 2019-May |
PublicationDateYYYYMMDD | 2019-05-01 |
PublicationDate_xml | – month: 05 year: 2019 text: 2019-May |
PublicationDecade | 2010 |
PublicationTitle | 2019 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) |
PublicationTitleAbbrev | IPDPSW |
PublicationYear | 2019 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
SSID | ssj0002685090 |
Score | 1.7426807 |
Snippet | Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 95 |
SubjectTerms | Biological neural networks Convolution Deep Larning Deep learning Field programmable gate arrays FPGA GPU Hardware Measurement Optimization Reconfigurable Computing |
Title | Deep Learning in Reconfigurable Hardware: A Survey |
URI | https://ieeexplore.ieee.org/document/8778332 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NSwMxEA21J08qrfhNDh7ddrNJk403sZYqVAq12FvZ2UxKEday7Cr660122yriwVvIJZmE5E1m5r0QciljA-gcg4BbaQNhIAl0giZwnjImIaYKKpLY6FEOp-Jh1ps1yNWWC4OIVfEZdnyzyuWb17T0obJurFTMubtwd5TSNVdrG0-JZOywb5O4ZKHu3o_748mzr97ykpSVfMKPD1Qq_BjskdFm5Lps5KVTFtBJP3-JMv53avuk_c3Uo-MtBh2QBmYtEvURV3StnLqgy4z6R2Zml4sy90wp6tP170mO1_SGTsr8DT_aZDq4e7odBuu_EYKlA_wiMD3DFechcgkiZcZKEQoGlhmBYLVhqY4RFHIbCky1joCrUCIYliALgfFD0sxeMzwiVDNumEAmgFkhItQAoBIjjfP03ImNjknL2ztf1fIX87WpJ393n5Jdv-J1TeAZaRZ5iecOtwu4qDbsCxXBmJc |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PT8IwFG4IHvSkBoy_7cGjg3Ut7erNiAQVCAkQuZF1fSXEZJBl0-hfb7sBGuPBW9NL-9q033t97_uK0DUPtQLrGHjUcOMxrSJPRqA96ylD5EMsVEES6w94d8Kepq1pBd1suTAAUBSfQcM1i1y-Xsa5eyprhkKElNoLd6dlowpRsrW2LyoBDy36bVKXxJfNx2F7OHpx9VtOlLIQUPjxhUqBIJ191N-MXRaOvDbyTDXiz1-yjP-d3AGqf3P18HCLQoeoAkkNBW2AFV5rp87xIsEuzEzMYp6njiuFXcL-PUrhFt_hUZ6-wUcdTToP4_uut_4dwVtYyM883dJUUOoD5YrFRBvOfEaUIZqBMlKTWIagBFDjM4ilDBQVPgelSQTEV4QeoWqyTOAYYUmoJgwIU8QwFoBUSolIc219PXtmgxNUc_bOVqUAxmxt6unf3Vdotzvu92a9x8HzGdpzq19WCJ6japbmcGFRPFOXxeZ9AU7om-g |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2019+IEEE+International+Parallel+and+Distributed+Processing+Symposium+Workshops+%28IPDPSW%29&rft.atitle=Deep+Learning+in+Reconfigurable+Hardware%3A+A+Survey&rft.au=Dias%2C+Mauricio+A.&rft.au=Ferreira%2C+Daniel+A.P.&rft.date=2019-05-01&rft.pub=IEEE&rft.spage=95&rft.epage=98&rft_id=info:doi/10.1109%2FIPDPSW.2019.00026&rft.externalDocID=8778332 |