Saliency-based YOLO for single target detection
Autor: | Jiang-she Zhang, Jun-ying Hu, C.-J. Richard Shi |
---|---|
Rok vydání: | 2021 |
Předmět: |
Artificial neural network
business.industry Computer science 02 engineering and technology Grid cell Object (computer science) Object detection Image (mathematics) Human-Computer Interaction Artificial Intelligence Hardware and Architecture Salient 020204 information systems 0202 electrical engineering electronic engineering information engineering Detection performance Saliency map Computer vision Artificial intelligence business Software Information Systems |
Zdroj: | Knowledge and Information Systems. 63:717-732 |
ISSN: | 0219-3116 0219-1377 |
DOI: | 10.1007/s10115-020-01538-0 |
Popis: | At present, You only look once (YOLO) is the fastest real-time object detection system based on a unified deep neural network. During training, YOLO divides the input image to $$S \times S $$ gird cells and the only one grid cell that contains the center of an object, takes charge of detecting that object. It is not sure that the cell corresponding to the center of the object is the best choice to detect the object. In this paper, inspired by the visual saliency mechanism we introduce the saliency map to YOLO to develop YOLO3-SM method, where saliency map selects the grid cell containing the most salient part of the object to detect the object. The experimental results on two data sets show that the prediction box of YOLO3-SM obtains the lager IOU value, which demonstrates that compared with YOLO3 , the YOLO3-SM selects the cell that is more suitable to detect the object . In addition, YOLO3-SM gets the highest mAP that the other three state-of-the-art object detection methods on the two data sets, which shows that introducing the saliency map to YOLO can improve the detection performance. |
Databáze: | OpenAIRE |
Externí odkaz: |