Cross-Modality Binary Code Learning via Fusion Similarity Hashing
Citations Over TimeTop 10% of 2017 papers
Abstract
Binary code learning has been emerging topic in large-scale cross-modality retrieval recently. It aims to map features from multiple modalities into a common Hamming space, where the cross-modality similarity can be approximated efficiently via Hamming distance. To this end, most existing works learn binary codes directly from data instances in multiple modalities, which preserve both intra-and inter-modal similarities respectively. Few methods consider to preserve the fusion similarity among multi-modal instances instead, which can explicitly capture their heterogeneous correlation in cross-modality retrieval. In this paper, we propose a hashing scheme, termed Fusion Similarity Hashing (FSH), which explicitly embeds the graph-based fusion similarity across modalities into a common Hamming space. Inspired by the fusion by diffusion, our core idea is to construct an undirected asymmetric graph to model the fusion similarity among different modalities, upon which a graph hashing scheme with alternating optimization is introduced to learn binary codes that embeds such fusion similarity. Quantitative evaluations on three widely used benchmarks, i.e., UCI Handwritten Digit, MIR-Flickr25K and NUS-WIDE, demonstrate that the proposed FSH approach can achieve superior performance over the state-of-the-art methods.
Related Papers
- → A Fast Approximate Nearest Neighbor Search Algorithm in the Hamming Space(2012)53 cited
- → Separated Variational Hashing Networks for Cross-Modal Retrieval(2019)34 cited
- → Efficient Processing of Hamming-Distance-Based Similarity-Search Queries Over MapReduce(2015)14 cited
- → Methods for multidimensional patterns recognition in Hamming space(2017)2 cited
- LOGICAL MEANING OF HAMMING SPHERE AND ITS GENERAL JUDGEMENT METHOD IN BINARY NEURAL NETWORKS(2002)