IEEE Access (Jan 2024)
LAYN: Lightweight Multi-Scale Attention YOLOv8 Network for Small Object Detection
Abstract
Currently, with the widespread application of embedded technology and the continuous improvement of computational power in mobile terminals, the efficient deployment of algorithms on embedded devices, while maintaining high accuracy and minimizing model size, has become a research hotspot. This paper addresses the challenges of deploying the YOLOv8 algorithm on embedded devices and proposes a novel lightweight object detection algorithm focusing on small object detection. We optimize the model through two key strategies, aiming to achieve lightweight deployment and improve the accuracy of small object detection. Firstly, GhostNet is introduced as the backbone network for YOLOv8 in order to achieve lightweight deployment. By using some cost effective operations to generate redundant feature maps, we not only reduce the number of model parameters while ensuring better detection results, but also improve the speed of the model. Secondly, a new multi-scale attention module is designed to enhance the network’s acquisition of crucial information for small targets, which includes a multi-scale fusion attention mechanism and the Soft-NMS algorithm. The multi-scale fusion attention mechanism captures key features of discriminative small targets in the feature map tensor from both spatial and channel dimensions, suppressing non-key information, reducing the impact of complex and unimportant information in the image, enhancing the network model’s learning ability for important features of small targets. The Soft-NMS method improves accuracy by significantly reduces false positives in the detection results. To validate the performance of our proposed method, we conducted validation experiments on the PASCAL VOC dataset and evaluated the model’s generalization ability on the MS COCO dataset. The experiments results demonstrate that our model achieves a significant improvement in small object detection, with a 5.41% increase in detection accuracy compared to the existing YOLOv8. Meanwhile, FLOPs are reduced by 49.62%, and the number of model parameters is reduced by 48.66%. These results fully confirm the effectiveness of our innovative method in achieving both lightweight deployment and significant efficacy in small object detection tasks.
Keywords