An Efficient Multi-Scale Focusing Attention Network for Person Re-Identification

Autor: Jihui Xu, Xiaoyu Hou, Wei Huang, Yongying Li, Huaiyu Xu, Kunlin Zhang, Ruidan Su
Jazyk: angličtina
Rok vydání: 2021
Předmět:
Computational complexity theory
Computer science
02 engineering and technology
lcsh:Technology
01 natural sciences
lcsh:Chemistry
lightweight network
0103 physical sciences
Convergence (routing)
0202 electrical engineering
electronic engineering
information engineering

General Materials Science
Latency (engineering)
lcsh:QH301-705.5
Instrumentation
Block (data storage)
010302 applied physics
Fluid Flow and Transfer Processes
multi-scale feature learning
person ReID
lcsh:T
Process Chemistry and Technology
General Engineering
Grid
softmax and TriHard loss
lcsh:QC1-999
Computer Science Applications
lcsh:Biology (General)
lcsh:QD1-999
Computer engineering
lcsh:TA1-2040
Softmax function
Focusing attention
020201 artificial intelligence & image processing
lcsh:Engineering (General). Civil engineering (General)
attention mechanism
lcsh:Physics
Communication channel
Zdroj: Applied Sciences
Volume 11
Issue 5
Applied Sciences, Vol 11, Iss 2010, p 2010 (2021)
ISSN: 2076-3417
DOI: 10.3390/app11052010
Popis: The multi-scale lightweight network and attention mechanism recently attracted attention in person re-identification (ReID) as it is capable of improving the model’s ability to process information with low computational cost. However, state-of-the-art methods mostly concentrate on the spatial attention and big block channel attention model with high computational complexity while rarely investigate the inside block attention with the lightweight network, which cannot meet the requirements of high efficiency and low latency in the actual ReID system. In this paper, a novel lightweight person ReID model is designed firstly, called Multi-Scale Focusing Attention Network (MSFANet), to capture robust and elaborate multi-scale ReID features, which have fewer float-computing and higher performance. MSFANet is achieved by designing a multi-branch depthwise separable convolution module, combining with an inside block attention module, to extract and fuse multi-scale features independently. In addition, we design a multi-stage backbone with the ‘1-2-3’ form, which can significantly reduce computational cost. Furthermore, the MSFANet is exceptionally lightweight and can be embedded in a ReID framework flexibly. Secondly, an efficient loss function combining softmax loss and TriHard loss, based on the proposed optimal data augmentation method, is designed for faster convergence and better model generalization ability. Finally, the experimental results on two big ReID datasets (Market1501 and DukeMTMC) and two small ReID datasets (VIPeR, GRID) show that the proposed MSFANet achieves the best mAP performance and the lowest computational complexity compared with state-of-the-art methods, which are increasing by 2.3% and decreasing by 18.2%, respectively.
Databáze: OpenAIRE