论文标题
feffnet:用于大规模细粒图像检索的统一哈希网络
ExchNet: A Unified Hashing Network for Large-Scale Fine-Grained Image Retrieval
论文作者
论文摘要
从大规模的细粒数据集中检索内容相关图像可能会遭受不可降低的查询速度和高度冗余的存储成本,这是由于高维的实价嵌入嵌入,旨在区分细粒物体的细微视觉差异。在本文中,我们研究了新颖的细粒度散列主题,以生成用于细粒图像的紧凑型二元代码,利用哈希学习的搜索和存储效率来减轻上述问题。具体而言,我们提出了一个统一的端到端可训练网络,称为“备用”。基于注意机制和提出的注意力约束,它可以首先获得局部和全局特征,以分别表示对象部分和整个细粒对象。此外,为了确保这些零件级特征在图像之间的歧视能力和语义含义的一致性,我们通过执行功能交换操作来设计局部特征对齐方法。后来,采用了一种替代学习算法来优化整个口气,然后生成最终的二进制哈希码。通过广泛的实验验证,我们的提案始终在五个细粒数据集上胜过最先进的通用哈希方法,这表明了我们的有效性。此外,与其他近似最近的邻居方法相比,Cheffnet实现了最佳的加速和存储降低,从而揭示了其效率和实用性。
Retrieving content relevant images from a large-scale fine-grained dataset could suffer from intolerably slow query speed and highly redundant storage cost, due to high-dimensional real-valued embeddings which aim to distinguish subtle visual differences of fine-grained objects. In this paper, we study the novel fine-grained hashing topic to generate compact binary codes for fine-grained images, leveraging the search and storage efficiency of hash learning to alleviate the aforementioned problems. Specifically, we propose a unified end-to-end trainable network, termed as ExchNet. Based on attention mechanisms and proposed attention constraints, it can firstly obtain both local and global features to represent object parts and whole fine-grained objects, respectively. Furthermore, to ensure the discriminative ability and semantic meaning's consistency of these part-level features across images, we design a local feature alignment approach by performing a feature exchanging operation. Later, an alternative learning algorithm is employed to optimize the whole ExchNet and then generate the final binary hash codes. Validated by extensive experiments, our proposal consistently outperforms state-of-the-art generic hashing methods on five fine-grained datasets, which shows our effectiveness. Moreover, compared with other approximate nearest neighbor methods, ExchNet achieves the best speed-up and storage reduction, revealing its efficiency and practicality.