Deep Learning in Reconfigurable Hardware: A Survey

Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related to the size and complexity of the networks and training algorithms. However, these structures are considerably resource-consuming and demand...

Full description

Saved in:
Bibliographic Details
Published in2019 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) pp. 95 - 98
Main Authors Dias, Mauricio A., Ferreira, Daniel A.P.
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.05.2019
Subjects
Online AccessGet full text
DOI10.1109/IPDPSW.2019.00026

Cover

Abstract Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related to the size and complexity of the networks and training algorithms. However, these structures are considerably resource-consuming and demand extra effort to be used on embedded systems. Researchers have chosen different alternatives to execute deep learning algorithms as servers, clusters, GPUs and FPGAs. Specific hardware structures have been designed to solve these problems in different platforms such as FPGAs and even ASICs. Although there are surveys about this subject, they do not present a clear criteria to real implementations, detailed explanation of design techniques and applied metrics. Presenting a different approach, this work analyzes reconfigurable hardware based structures designed to optimize deep learning algorithms. This work includes results of hardware accelerators implemented and tested using commercially available development boards avoiding simulation-based results.Results of the proposed analysis showed that a lot of effort is directed to this subject but presented results are far from what is expected for the performance of a hardware structure for deep learning.
AbstractList Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related to the size and complexity of the networks and training algorithms. However, these structures are considerably resource-consuming and demand extra effort to be used on embedded systems. Researchers have chosen different alternatives to execute deep learning algorithms as servers, clusters, GPUs and FPGAs. Specific hardware structures have been designed to solve these problems in different platforms such as FPGAs and even ASICs. Although there are surveys about this subject, they do not present a clear criteria to real implementations, detailed explanation of design techniques and applied metrics. Presenting a different approach, this work analyzes reconfigurable hardware based structures designed to optimize deep learning algorithms. This work includes results of hardware accelerators implemented and tested using commercially available development boards avoiding simulation-based results.Results of the proposed analysis showed that a lot of effort is directed to this subject but presented results are far from what is expected for the performance of a hardware structure for deep learning.
Author Dias, Mauricio A.
Ferreira, Daniel A.P.
Author_xml – sequence: 1
  givenname: Mauricio A.
  surname: Dias
  fullname: Dias, Mauricio A.
  email: macdias@fho.edu.br
  organization: University Center of Heminio Ometto Foundation
– sequence: 2
  givenname: Daniel A.P.
  surname: Ferreira
  fullname: Ferreira, Daniel A.P.
  email: danielferreira@fho.edu.br
  organization: University Center of Heminio Ometto Foundation
BookMark eNotjMFKAzEUACMoaGs_QLzkB7a-l7e72XgrrdrCgsUqHkuyeSmRmpasVfr3inoamIEZiNO0SyzEFcIYEczNYjlbrl7HCtCMAUDVJ2KAWjVI1U8-F6O-f_v1TQUGLoSaMe9lyzanmDYyJvnE3S6FuDlk67Ys5zb7L5v5Vk7k6pA_-XgpzoLd9jz651C83N89T-dF-_iwmE7aIqKuPgpfedJEwFS7skMf6hJKdAF9yS4Yj51p2GmmACV3xihHGmp2Hi0jOKShuP77RmZe73N8t_m4brRuiBR9A-pxRE4
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/IPDPSW.2019.00026
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 1728135109
9781728135106
EndPage 98
ExternalDocumentID 8778332
Genre orig-research
GroupedDBID 6IE
6IF
6IL
6IN
AAJGR
AAWTH
ABLEC
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
IEGSK
OCL
RIB
RIC
RIE
RIL
ID FETCH-LOGICAL-i175t-d5d37330e36b4c1df64041bf1d4ebf9d1c98eb7e3f04ec992b3706ebd1ae10b13
IEDL.DBID RIE
IngestDate Wed Aug 27 02:40:28 EDT 2025
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i175t-d5d37330e36b4c1df64041bf1d4ebf9d1c98eb7e3f04ec992b3706ebd1ae10b13
PageCount 4
ParticipantIDs ieee_primary_8778332
PublicationCentury 2000
PublicationDate 2019-May
PublicationDateYYYYMMDD 2019-05-01
PublicationDate_xml – month: 05
  year: 2019
  text: 2019-May
PublicationDecade 2010
PublicationTitle 2019 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW)
PublicationTitleAbbrev IPDPSW
PublicationYear 2019
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0002685090
Score 1.7426807
Snippet Deep Learning has been applied successfully to solve complex problems that involves analysis of large data sets and these good results can be directly related...
SourceID ieee
SourceType Publisher
StartPage 95
SubjectTerms Biological neural networks
Convolution
Deep Larning
Deep learning
Field programmable gate arrays
FPGA
GPU
Hardware
Measurement
Optimization
Reconfigurable Computing
Title Deep Learning in Reconfigurable Hardware: A Survey
URI https://ieeexplore.ieee.org/document/8778332
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NSwMxEA21J08qrfhNDh7ddrNJk403sZYqVAq12FvZ2UxKEday7Cr660122yriwVvIJZmE5E1m5r0QciljA-gcg4BbaQNhIAl0giZwnjImIaYKKpLY6FEOp-Jh1ps1yNWWC4OIVfEZdnyzyuWb17T0obJurFTMubtwd5TSNVdrG0-JZOywb5O4ZKHu3o_748mzr97ykpSVfMKPD1Qq_BjskdFm5Lps5KVTFtBJP3-JMv53avuk_c3Uo-MtBh2QBmYtEvURV3StnLqgy4z6R2Zml4sy90wp6tP170mO1_SGTsr8DT_aZDq4e7odBuu_EYKlA_wiMD3DFechcgkiZcZKEQoGlhmBYLVhqY4RFHIbCky1joCrUCIYliALgfFD0sxeMzwiVDNumEAmgFkhItQAoBIjjfP03ImNjknL2ztf1fIX87WpJ393n5Jdv-J1TeAZaRZ5iecOtwu4qDbsCxXBmJc
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PT8IwFG4IHvSkBoy_7cGjg3Ut7erNiAQVCAkQuZF1fSXEZJBl0-hfb7sBGuPBW9NL-9q033t97_uK0DUPtQLrGHjUcOMxrSJPRqA96ylD5EMsVEES6w94d8Kepq1pBd1suTAAUBSfQcM1i1y-Xsa5eyprhkKElNoLd6dlowpRsrW2LyoBDy36bVKXxJfNx2F7OHpx9VtOlLIQUPjxhUqBIJ191N-MXRaOvDbyTDXiz1-yjP-d3AGqf3P18HCLQoeoAkkNBW2AFV5rp87xIsEuzEzMYp6njiuFXcL-PUrhFt_hUZ6-wUcdTToP4_uut_4dwVtYyM883dJUUOoD5YrFRBvOfEaUIZqBMlKTWIagBFDjM4ilDBQVPgelSQTEV4QeoWqyTOAYYUmoJgwIU8QwFoBUSolIc219PXtmgxNUc_bOVqUAxmxt6unf3Vdotzvu92a9x8HzGdpzq19WCJ6japbmcGFRPFOXxeZ9AU7om-g
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2019+IEEE+International+Parallel+and+Distributed+Processing+Symposium+Workshops+%28IPDPSW%29&rft.atitle=Deep+Learning+in+Reconfigurable+Hardware%3A+A+Survey&rft.au=Dias%2C+Mauricio+A.&rft.au=Ferreira%2C+Daniel+A.P.&rft.date=2019-05-01&rft.pub=IEEE&rft.spage=95&rft.epage=98&rft_id=info:doi/10.1109%2FIPDPSW.2019.00026&rft.externalDocID=8778332