Paper ID | MLR-APPL-IVASR-3.4 | ||
Paper Title | APNET: ATTRIBUTE PARSING NETWORK FOR PERSON RE-IDENTIFICATION | ||
Authors | Chiat Pin Tay, Kim Hui Yap, Nanyang Technological University, Singapore | ||
Session | MLR-APPL-IVASR-3: Machine learning for image and video analysis, synthesis, and retrieval 3 | ||
Location | Area E | ||
Session Time: | Tuesday, 21 September, 08:00 - 09:30 | ||
Presentation Time: | Tuesday, 21 September, 08:00 - 09:30 | ||
Presentation | Poster | ||
Topic | Applications of Machine Learning: Machine learning for image & video analysis, synthesis, and retrieval | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Most person re-identification methods rely solely on the pedestrian identity for learning. Person attributes, such as gender, clothing colors, carried bags, etc, are however seldom used. These attributes are highly identity-related and should be capitalized fully. Thus, we propose Attribute Parsing Network (APNet), an architecture designed for both image and person attribute learning and retrievals. To further enhance the re-id performance, we propose to leverage saliency maps and human parsing to boost the foreground features, which when trained together with the global and local networks, resulted in more generic and robust encoded representations. This proposed method achieved state-of-the-art accuracy performance on both Market1501 (87.3\% mAP and 95.2\% Rank1) and DukeMTMC-reID (78.8\% mAP and 89.2\% Rank 1) datasets. |