HashNet: Deep Learning to Hash by Continuation

Learning to hash has been widely applied to approximate nearest neighbor search for large-scale multimedia retrieval, due to its computation efficiency and retrieval quality. Deep learning to hash, which improves retrieval quality by end-to-end representation learning and hash encoding, has received...

Full description

Saved in:
Bibliographic Details
Published inProceedings / IEEE International Conference on Computer Vision pp. 5609 - 5618
Main Authors Zhangjie Cao, Mingsheng Long, Jianmin Wang, Yu, Philip S.
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.10.2017
Subjects
Online AccessGet full text
ISSN2380-7504
DOI10.1109/ICCV.2017.598

Cover

Abstract Learning to hash has been widely applied to approximate nearest neighbor search for large-scale multimedia retrieval, due to its computation efficiency and retrieval quality. Deep learning to hash, which improves retrieval quality by end-to-end representation learning and hash encoding, has received increasing attention recently. Subject to the ill-posed gradient difficulty in the optimization with sign activations, existing deep learning to hash methods need to first learn continuous representations and then generate binary hash codes in a separated binarization step, which suffer from substantial loss of retrieval quality. This work presents HashNet, a novel deep architecture for deep learning to hash by continuation method with convergence guarantees, which learns exactly binary hash codes from imbalanced similarity data. The key idea is to attack the ill-posed gradient problem in optimizing deep networks with non-smooth binary activations by continuation method, in which we begin from learning an easier network with smoothed activation function and let it evolve during the training, until it eventually goes back to being the original, difficult to optimize, deep network with the sign activation function. Comprehensive empirical evidence shows that HashNet can generate exactly binary hash codes and yield state-of-the-art multimedia retrieval performance on standard benchmarks.
AbstractList Learning to hash has been widely applied to approximate nearest neighbor search for large-scale multimedia retrieval, due to its computation efficiency and retrieval quality. Deep learning to hash, which improves retrieval quality by end-to-end representation learning and hash encoding, has received increasing attention recently. Subject to the ill-posed gradient difficulty in the optimization with sign activations, existing deep learning to hash methods need to first learn continuous representations and then generate binary hash codes in a separated binarization step, which suffer from substantial loss of retrieval quality. This work presents HashNet, a novel deep architecture for deep learning to hash by continuation method with convergence guarantees, which learns exactly binary hash codes from imbalanced similarity data. The key idea is to attack the ill-posed gradient problem in optimizing deep networks with non-smooth binary activations by continuation method, in which we begin from learning an easier network with smoothed activation function and let it evolve during the training, until it eventually goes back to being the original, difficult to optimize, deep network with the sign activation function. Comprehensive empirical evidence shows that HashNet can generate exactly binary hash codes and yield state-of-the-art multimedia retrieval performance on standard benchmarks.
Author Zhangjie Cao
Yu, Philip S.
Mingsheng Long
Jianmin Wang
Author_xml – sequence: 1
  surname: Zhangjie Cao
  fullname: Zhangjie Cao
  email: caozhangjie14@gmail.com
  organization: KLiss, Tsinghua Univ., Beijing, China
– sequence: 2
  surname: Mingsheng Long
  fullname: Mingsheng Long
  email: mingsheng@tsinghua.edu.cn
  organization: KLiss, Tsinghua Univ., Beijing, China
– sequence: 3
  surname: Jianmin Wang
  fullname: Jianmin Wang
  email: jimwang@tsinghua.edu.cn
  organization: KLiss, Tsinghua Univ., Beijing, China
– sequence: 4
  givenname: Philip S.
  surname: Yu
  fullname: Yu, Philip S.
  email: psyu@uic.edu
  organization: KLiss, Tsinghua Univ., Beijing, China
BookMark eNotjk1PwzAQRA0Cibb0yImL_0CC12vHa24ofLRSBBfgWjnJBozAqZJw6L8nCC4z0hvpaZbiJPWJhbgAlQMof7Uty9dcK3C59XQk1t4RWKQCFGp_LBYaSWXOKnMmluP4oRR6TcVC5Jswvj_ydC1vmfey4jCkmN7k1MvfRdYHWfZpiuk7TLFP5-K0C58jr_97JV7u757LTVY9PWzLmyqL4OyU-SIghKatGxsabhnb-Q0xtXbmpKmew3htjNMFoWqAHaLvjKuhA1N3uBKXf97IzLv9EL_CcNiRRkeFwh9htkK4
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/ICCV.2017.598
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Applied Sciences
EISBN 9781538610329
1538610329
EISSN 2380-7504
EndPage 5618
ExternalDocumentID 8237860
Genre orig-research
GroupedDBID 29O
6IE
6IF
6IH
6IK
6IL
6IM
6IN
AAJGR
AAWTH
ACGFS
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IPLJI
M43
OCL
RIE
RIL
RIO
RNS
ID FETCH-LOGICAL-i175t-96a31acdbc5acede3d8158e8d5a31828b18249244726830c1e7339f47b1f14bf3
IEDL.DBID RIE
IngestDate Wed Aug 27 02:42:16 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i175t-96a31acdbc5acede3d8158e8d5a31828b18249244726830c1e7339f47b1f14bf3
PageCount 10
ParticipantIDs ieee_primary_8237860
PublicationCentury 2000
PublicationDate 2017-Oct.
PublicationDateYYYYMMDD 2017-10-01
PublicationDate_xml – month: 10
  year: 2017
  text: 2017-Oct.
PublicationDecade 2010
PublicationTitle Proceedings / IEEE International Conference on Computer Vision
PublicationTitleAbbrev ICCV
PublicationYear 2017
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0039286
Score 2.5912373
Snippet Learning to hash has been widely applied to approximate nearest neighbor search for large-scale multimedia retrieval, due to its computation efficiency and...
SourceID ieee
SourceType Publisher
StartPage 5609
SubjectTerms Binary codes
Encoding
Machine learning
Optimization
Quantization (signal)
Semantics
Training
Title HashNet: Deep Learning to Hash by Continuation
URI https://ieeexplore.ieee.org/document/8237860
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PT8MgFCZzJ09TN-PvcPBou1KgBa_TZZps8eDMbkuBhy4m3eLag_71Qttt0XjwAgQO8CDw8eC97yF0LYVJiZIyyCjRAaPWRwPMeBBxrW2U2Nhy7-88niSjKXuc8VkL3Wx9YQCgMj6D0Berv3yz1KV_Kut7YhWROAV9z2W1r9bm1HUwL5Idh2b_YTB48YZbacil-BE5pQKOYQeNN13W9iLvYVmoUH_9YmP875gOUG_nooeftuBziFqQH6FOc6fEzY5dd1E4ytZvEyhu8R3ACjd0qq-4WGLfgtUn9gRVi7ym_O6h6fD-eTAKmhgJwcIBfxHIxE1ypo3SPNNggBpBuABhuKt32pRyCXM6FkvjRNBIE0gplZaliljClKXHqJ0vczhBWIpYx4ZQnmYZU2A9MVVMEmm0MHHE1CnqevHnq5oGY95IfvZ39Tna99Nf271doHbxUcKlw-9CXVUL9w3eX5oQ
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3NT8IwFG8IHvSECsZve_Doxrq2W-sVJUOBeADDjaxfSkgGkXHQv952GxCNBy9N0x7a15f219e-93sA3HKmYiQ491KMpEewcdkAU-oFVEoTRCY01MU7D4ZRMiZPEzqpgbttLIzWunA-076rFn_5aiHX7qms7YhVWGQN9D1KCKFltNbm3LVAz6Idi2a71-m8Otet2Kec_cidUkBHtwEGm0FLj5G5v86FL79-8TH-d1aHoLUL0oMvW_g5AjWdHYNGdauE1Z5dNYGfpKv3oc7v4YPWS1gRqr7BfAFdDxSf0FFUzbKS9LsFxt3HUSfxqiwJ3sxCf-7xyC5zKpWQNJVaaawYokwzRW27taeELYi1skgcRgwHEukYY25ILJBBRBh8AurZItOnAHIWylAhTOM0JUIbR00VoogryVQYEHEGmk786bIkwphWkp__3XwD9pPRoD_t94bPF-DAqaL0grsE9fxjra8smufiulDiNyQAnV0
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=Proceedings+%2F+IEEE+International+Conference+on+Computer+Vision&rft.atitle=HashNet%3A+Deep+Learning+to+Hash+by+Continuation&rft.au=Zhangjie+Cao&rft.au=Mingsheng+Long&rft.au=Jianmin+Wang&rft.au=Yu%2C+Philip+S.&rft.date=2017-10-01&rft.pub=IEEE&rft.eissn=2380-7504&rft.spage=5609&rft.epage=5618&rft_id=info:doi/10.1109%2FICCV.2017.598&rft.externalDocID=8237860