MCR-YOLO model for underwater target detection based on multi-color spatial features

Peiyu Liu, Weiwei Xing, Yixuan Ma

Optoelectronics Letters ›› 2024, Vol. 20 ›› Issue (5) : 313-320. DOI: 10.1007/s11801-024-3248-5
Article

MCR-YOLO model for underwater target detection based on multi-color spatial features

Author information +
History +

Abstract

Within the fields of underwater robotics and ocean information processing, computer vision-based underwater target detection is an important area of research. Underwater target detection is made more difficult by a number of problems with underwater imagery, such as low contrast, color distortion, fuzzy texture features, and noise interference, which are caused by the limitations of the unique underwater imaging environment. In order to solve the above challenges, this paper proposes a multi-color space residual you only look once (MCR-YOLO) model for underwater target detection. First, the RGB image is converted into YCbCr space, and the brightness channel Y is used to extract the non-color features of color-biased images based on improved ResNet50. Then, the output features of three scales are combined between adjacent scales to exchange information. At the same time, the image features integrated with low-frequency information are obtained via the low-frequency feature extraction branch and the three-channel RGB image, and the features from the three scales of the two branches are fused at the corresponding scales. Finally, multi-scale fusion and target detection are accomplished utilizing the path aggregation network (PANet) framework. Experiments on relevant datasets demonstrate that the method can improve feature extraction of critical targets in underwater environments and achieve good detection accuracy.

Cite this article

Download citation ▾
Peiyu Liu, Weiwei Xing, Yixuan Ma. MCR-YOLO model for underwater target detection based on multi-color spatial features. Optoelectronics Letters, 2024, 20(5): 313‒320 https://doi.org/10.1007/s11801-024-3248-5

References

[[1]]
Zhang S, Wang T, Dong J, et al.. Underwater image enhancement via extended multi-scale retinex. Neuro computing, 2017, 245: 1-9 [J]
[[2]]
Ren S, He K, Girshick R, et al.. Faster R-CNN: towards real-time object detection with region proposal networks. IEEE transactions on pattern analysis and machine intelligence, 2017, 39(6): 1137-1149, J]
CrossRef Google scholar
[[3]]
BOCHKOVSKIY A, WANG C Y, LIAO H Y M. YOLOv4: optimal speed and accuracy of object detec-tion[EB/OL]. (2020-04-23) [2022-10-17]. https://arxiv.org/abs/2004.10934.
[[4]]
Ancuti C O, Ancuti C, Vleeschouwer C D, et al.. Color balance and fusion for underwater image enhancement. IEEE transactions on image processing, 2018, 27(1): 379-393, J]
CrossRef Google scholar
[[5]]
Lin W H, Zhong J X, Liu S, et al.. RoIMix: proposal-fusion among multiple images for underwater object detection. IEEE International Conference on Acoustics, Speech and Signal Processing, May 4–8, 2020, Barcelona, Spain, 2020 New York IEEE 2588-2592 [C]
[[6]]
Jian M, Liu X, Luo H, et al.. Underwater image processing and analysis: a review. Signal processing: image communication, 2021, 91: 116088 [J]
[[7]]
CAI X, JIANG N, CHEN W, et al. CURE-Net: a cascaded deep network for underwater image enhance-ment[J]. IEEE journal of oceanic engineering, 2023.
[[8]]
Li Y, Ruan R, Mi Z, et al.. An underwater image restoration based on global polarization effects of underwater scene. Optics and lasers in engineering, 2023, 165: 107550, J]
CrossRef Google scholar
[[9]]
Fan B, Chen W, Cong Y, et al.. Dual refinement underwater object detection network. European Conference on Computer Vision, August 23–28, 2020, Glasgow, UK, 2020 Cham Springer 275-291 [C]
[[10]]
LIU H, SONG P, DING R. WQT and DG-YOLO: towards domain generalization in underwater object de-tection[EB/OL]. (2020-04-14) [2022-10-17]. http://arxiv.org/abs/2004.06333.
[[11]]
WANG Z, LIU C, WANG S, et al. UDD: an underwater open-sea farm object detection dataset for underwater robot picking[EB/OL]. (2021-07-28) [2022-10-17]. https://arxiv.org/abs/2003.01446v1.
[[12]]
Li Y, Ruan R, Mi Z, et al.. An underwater image restoration based on global polarization effects of underwater scene. Optics and lasers in engineering, 2023, 165: 107550, J]
CrossRef Google scholar
[[13]]
Zhang Y, Li X S, Sun Y M, et al.. Underwater object detection algorithm based on channel attention and feature fusion. Journal of Northwestern Polytechnical University, 2022, 40(2): 433-441, J]
CrossRef Google scholar
[[14]]
Yang Y, Yu H P, Zhao G L. A fast algorithm for YCbCr to RGB conversion. IEEE transactions on consumer electronics, 2007, 53(4): 1490-1493, J]
CrossRef Google scholar
[[15]]
Premal C E, Vinsley S S. Image processing based forest fire detection using YCbCr colour model. 2014 International Conference on Circuits, Power and Computing Technologies (ICCPCT-2014), March 20–21, 2014, Nagercoil, India, 2014 New York IEEE 1229-1237 [C]
[[16]]
Chen Y, Fan H, Xu B, et al.. Drop an octave: reducing spatial redundancy in convolutional neural networks with octave convolution. Proceedings of the IEEE/CVF International Conference on Computer Vision, October 27–November 2, 2019, Seoul, Korea (South), 2019 New York IEEE 3435-3444 [C]
[[17]]
PENG L, ZHU C, BIAN L. U-shape transformer for underwater image enhancement[EB/OL]. (2021-11-23) [2022-10-17]. http://arxiv.org/abs/2111.11843.
[[18]]
Mahasin M, Dewi I A. Comparison of CSPDark-Net53, CSPResNeXt-50, and EfficientNet-B0 backbones on YOLOv4 as object detector. International journal of engineering, science and information technology, 2022, 2(3): 64-72, J]
CrossRef Google scholar
[[19]]
He K, Zhang X, Ren S, et al.. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 27–30, 2016, Las Vegas, USA, 2016 New York IEEE 770-778 [C]
[[20]]
Sethi R, Indu S. Fusion of underwater image enhancement and restoration. International journal of pattern recognition and artificial intelligence, 2020, 34(03): 2054007, J]
CrossRef Google scholar
[[21]]
Zhang Q, Da L, Zhang Y, et al.. Integrated neural networks based on feature fusion for underwater target recognition. Applied acoustics, 2021, 182: 108261, J]
CrossRef Google scholar
[[22]]
Yang H H, Huang K C, Chen W T. Laffnet: a lightweight adaptive feature fusion network for underwater image enhancement. 2021 IEEE International Conference on Robotics and Automation (ICRA), May 30–June 6, 2021, Xi’an, China, 2021 New York IEEE 685-692 [C]

Accesses

Citations

Detail

Sections
Recommended

/