Relation-aware aggregation network with auxiliary guidance for text-based person search.

Autor: Zeng, Pengpeng, Jing, Shuaiqi, Song, Jingkuan, Fan, Kaixuan, Li, Xiangpeng, We, Liansuo, Guo, Yuan
Předmět:
Zdroj: World Wide Web; Jul2022, Vol. 25 Issue 4, p1565-1582, 18p
Abstrakt: In this paper, we propose a novel Relation-aware Aggregation Network with Auxiliary Guidance for text-based person search, namely RANAG. Existing works are still hard to capture the detailed appearance of a person and compute the similarity between images and texts. RANAN is designed to address the above problem from two aspects: relation-aware visual and additional auxiliary signal. Specifically, we introduce a Relation-aware Aggregation Network (RAN) that exploits the relation between the person and local objects. Then, we propose three auxiliary tasks to acquire additional knowledge of semantic representations. Each task has a respective objective: identifying the gender of the pedestrian in the image, distinguishing the images of the similar pedestrian, and aligning the semantic information between description and image. In addition, some data augmentation methods we explored can further improve the performance. Extensive experiments demonstrate that our model provides superior performance than the state-of-the-art methods on the CUHK-PEDES dataset. [ABSTRACT FROM AUTHOR]
Databáze: Complementary Index