摘要
We present a new method to generate efficient multi-level hashing codes for image retrieval based on the deep siamese conv01utional neural network (DSCNN). Con- ventional deep hashing methods trade off the capability of capturing highly complex and nonlinear semantic informa- tion of images against very compact hash codes, usually lead- ing to high retrieval efficiency but with deteriorated accuracy. We alleviate the restrictive compactness requirement of hash codes by extending them to a two-level hierarchical coding scheme, in which the first level aims to capture the high-level semantic information extracted by the deep network using a rich encoding strategy, while the subsequent level squeezes them to more global and compact codes. At running time, we adopt an attention-based mechanism to select some of its most essential bits specific to each query image for retrieval instead of using the full hash codes of the first level. The attention-based mechanism is based on the guides of hash codes generated by the second level, taking advantage of both local and global properties of deep features. Experimental re- suits on various popular datasets demonstrate the advantages of the proposed method compared to several state-of-the-art methods.
We present a new method to generate efficient multi-level hashing codes for image retrieval based on the deep siamese conv01utional neural network (DSCNN). Con- ventional deep hashing methods trade off the capability of capturing highly complex and nonlinear semantic informa- tion of images against very compact hash codes, usually lead- ing to high retrieval efficiency but with deteriorated accuracy. We alleviate the restrictive compactness requirement of hash codes by extending them to a two-level hierarchical coding scheme, in which the first level aims to capture the high-level semantic information extracted by the deep network using a rich encoding strategy, while the subsequent level squeezes them to more global and compact codes. At running time, we adopt an attention-based mechanism to select some of its most essential bits specific to each query image for retrieval instead of using the full hash codes of the first level. The attention-based mechanism is based on the guides of hash codes generated by the second level, taking advantage of both local and global properties of deep features. Experimental re- suits on various popular datasets demonstrate the advantages of the proposed method compared to several state-of-the-art methods.
基金
This work was partially supported by the National Natural Science Foundation of China (Grant Nos, 61373060 and 61672280) and Qing Lan Project.