Paper ID | ARS-4.11 | ||
Paper Title | GAT-ASSISTED DEEP HASHING FOR MULTI-LABEL IMAGE RETRIEVAL | ||
Authors | Jiachen Li, Zhi Li, Yanchun Ma, Qing Xie, Yongjian Liu, Wuhan University of Technology, China | ||
Session | ARS-4: Re-Identification and Retrieval | ||
Location | Area I | ||
Session Time: | Wednesday, 22 September, 08:00 - 09:30 | ||
Presentation Time: | Wednesday, 22 September, 08:00 - 09:30 | ||
Presentation | Poster | ||
Topic | Image and Video Analysis, Synthesis, and Retrieval: Image & Video Storage and Retrieval | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Multi-Label hash methods have achieved excellent performance in multi-label image retrieval, but how to leverage the semantic information of label to improve retrieval quality is still a challenge in this field. This paper proposes GAT-Assisted Deep Hashing (DHGAT). Our model uses Convolutional Neural Network (CNN) to extract image-level features, along with graph attention network (GAT) to extract label-level features. Assisted by GAT, DHGAT is able to pay more attention on the co-occurrence of label. In order to solve the problem of feature fusion, we propose Multi-modal Max-Pooling Bilinear (MMB) mechanism, which MMB fuses image-level feature and label-level feature to generate abundant semantic features, so that the model can output discriminative hash code. Extensive experiments demonstrate that the proposed method can generate hash codes which achieve better retrieval performance on two benchmark datasets, NUS-WIDE and MS-COCO. |