YOLO-FOA: A lightweight rotational target detection algorithm based on improved YOLO for optical fiber robot

Yingqi Wu , Jialong Chen , Xiuli Yu , Jian Li

Biomimetic Intelligence and Robotics ›› 2026, Vol. 6 ›› Issue (1) : 100273

PDF
Biomimetic Intelligence and Robotics ›› 2026, Vol. 6 ›› Issue (1) :100273 DOI: 10.1016/j.birob.2026.100273
Research Article
research-article
YOLO-FOA: A lightweight rotational target detection algorithm based on improved YOLO for optical fiber robot
Author information +
History +
PDF

Abstract

In fiber optic communication, as networks expand, precise detection and alignment of fiber optic adapters are crucial for enhancing system stability and transmission quality. Traditional target detection algorithms face two main issues in fiber optic adapter detection: inability to handle arbitrarily oriented targets and difficulty in efficient deployment on embedded devices. To tackle these issues, this paper introduces a lightweight rotating target detection algorithm, YOLO-FOA, for fiber optic communication scenarios. The algorithm is based on the YOLO model, which significantly reduces the computational and parametric quantities of the model by introducing Dynamic Head and Dynamic ATSS, and the C2f_MViTBv3, C2f_GhostBlockv2 modules and Angle DFL Loss are designed to improve the detection accuracy. In addition, the dynamic alignment correction mechanism can be effectively applied to intelligent calibration and real-time deviation correction in fiber optic communication networks. Experiments show YOLO-FOA achieves 97.1% detection accuracy on a self-constructed dataset, outperforming the baseline model by 1.3%, with a 4.5% reduction in parameters and 7.2% in computation. Suitable for embedded devices due to its high accuracy and low resource demands, YOLO-FOA offers a new approach to enhancing fiber optic communication system stability and transmission quality.

Keywords

Real-time target detection / Lightweight network / Dynamic alignment correction / Fiber optic adapter / Directional bracketing frame

Cite this article

Download citation ▾
Yingqi Wu, Jialong Chen, Xiuli Yu, Jian Li. YOLO-FOA: A lightweight rotational target detection algorithm based on improved YOLO for optical fiber robot. Biomimetic Intelligence and Robotics, 2026, 6(1): 100273 DOI:10.1016/j.birob.2026.100273

登录浏览全文

4963

注册一个新账户 忘记密码

CRediT authorship contribution statement

Yingqi Wu: Writing – review & editing, Writing – original draft. Jialong Chen: Writing – review & editing, Data curation. Xiuli Yu: Funding acquisition. Jian Li: Funding acquisition, Formal analysis.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

This work was supported by the interdisciplinary Team of Intelligent Elderly Care and Rehabilitation in the “Double first-class” Construction (530324004), and Talent Introduction Project (510224074) of Beijing University of Posts and Telecommunications, China .

Appendix A. Supplementary data

Supplementary material related to this article can be found online at https://doi.org/10.1016/j.birob.2026.100273.

References

[1]

Beijing Ruiqi Haodi Technology Co., Ltd , A method, system, device and electronic equipment for maintaining fiber channel services: CN202311843437.7 [P], (2024).

[2]

Wang Xu, Wu Yanxia, Zhang Xue, et al., Review on rotation target detection under computer vision, Comput. Sci. 50 (8) (2023) 79-92, https://doi.org/10.11896/jsjkx.221000148.

[3]

S. Yang, Z. Pei, F. Zhou, et al., Rotated faster r-cnn for oriented object detection in aerial images[C], Proceedings of the 2020 3rd International Conference on Robot Systems and Applications, 2020, pp. 35-39.

[4]

J. Ding, N. Xue, Y. Long, et al., Learning roi transformer for oriented object detection in aerial images[C], Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 2849-2858.

[5]

J. Ma, W. Shao, H. Ye, et al., Arbitrary-oriented scene text detection via rotation proposals, IEEE Trans. Multimed. 20 (11) (2018) 3111-3122.

[6]

W. Li, Y. Chen, K. Hu, et al., Oriented reppoints for aerial object detection[C], Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 1829-1838.

[7]

T.T. Fang, B. Liu, C. Chen, X.Y. Li, View-consistency network for weakly supervised oriented object detection in remote sensing images, Natl. Remote. Sens. Bull. 28 (12) (2024) 3213-3230, https://doi.org/10.11834/jrs.20243138.

[8]

Penglei Wang, Wenjing Hu, Jiping Zhang, et al., Enhanced Rotated Mask R-CNN for Chromosome Segmentation[C], in: 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society: EMBC 2021, Virtual Conference, 1-5 2021, 2021, pp. 2769-2772, [v.4].

[9]

Chao Ye, Jiangang Ding, Huaizhi Wang, et al., Enhanced security contraband detection through integration of attention mechanism in R3Det [J], J. Phys.:Conf. Ser., 2589 (1), (2023), 10.1088/1742-6596/2589/1/012002.

[10]

Liu Bingbing, Zhang Qinglin, Improved YOLOv5s algorithm for rotated target detection in remote sensing images, Laser J., 45 (12), (2024).

[11]

X. Yang, J. Yan, Z. Feng, et al., R3det: Refined single-stage detector with feature refinement for rotating object[c], Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, 2021, pp. 3163-3171.

[12]

P. Zhao, Z. Qu, Y. Bu, et al., Polardet: A fast, more precise detector for rotated target in aerial images, Int. J. Remote Sens. 42 (15) (2021) 5831-5861.

[13]

R. Girshick, J. Donahue, T. Darrell, et al., Rich feature hierarchies for accurate object detection and semantic segmentation [C], in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 580-587.

[14]

S. Ren, K. He, R. Girshick, et al., Faster r-cnn: Towards real-time object detection with region proposal networks, Advances in Neural Information Processing Systems, (2015), p. 28.

[15]

W. Liu, D. Anguelov, D. Erhan, et al., Ssd: Single shot multibox detector [C], Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, the Netherlands, October (2016) 11–14, Proceedings, Part I 14, Springer, 2016, pp. 21-37.

[16]

T.Y. Lin, P. Goyal, R. Girshick, et al., Focal loss for dense object detection [C], in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2980-2988.

[17]

J. Redmon, S. Divvala, R. Girshick, et al., You only look once: Unified, real-time object detection [C], in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 779-788.

[18]

T.Y. Lin, P. Dollár, R. Girshick, et al., Feature pyramid networks for object detection [C], in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2117-2125.

[19]

M. Tan, R. Pang, Q.-V. Le, Efficientdet: Scalable and efficient object detection [C], in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 10781-10790.

[20]

A. Dosovitskiy, L. Beyer, A. Kolesnikov, et al., An Image is Worth 16 ×16 Words: Transformers for Image Recognition at Scale, in: Proceedings of the International Conference on Learning Representations, ICLR, 2021.

[21]

N. Carion, F. Massa, G. Synnaeve, et al., End-to-end object detection with transformers [C], European Conference on Computer Vision, Springer, 2020, pp. 213-229.

[22]

X. Zhu, W. Su, L. Lu, et al., Deformable DETR: Deformable Transformers for End-to-End Object Detection, in: Proceedings of the International Conference on Learning Representations, ICLR, 2021.

[23]

Z. Liu, Y. Lin, Y. Cao, et al., Swin transformer: Hierarchical vision transformer using shifted win-dows [C], in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 10012-10022.

[24]

S. Fu, Q. Yang, Q. Mo, J. Yan, X. Wei, J. Meng, X. Xie, W.-S. Zheng, LLMDet: Learning strong open-vocabulary object detectors under the supervision of large language models, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR, 2025, pp. 14987-14997, https://doi.org/10.1109/CVPR52734.2025.01396.

[25]

Ruijiao Jin, Kun Wang, Minhao Liu, Xichao Teng, Zhang Li, Qifeng Yu, DETR with improved DeNoising training for multi-scale oriented object detection in optical remote sensing images (Invited), Laser Optoelectron. Prog. 61 (2) (2024) 0211023.

[26]

Y. Xu, M. Fu, Q. Wang, et al., Gliding vertex on the horizontal bounding box for multi-oriented object detection, IEEE Trans. Pattern Anal. Mach. Intell. 43 (4) (2020) 1452-1459.

[27]

H. Wei, Y. Zhang, Z. Chang, et al., Oriented objects as pairs of middle lines, ISPRS J. Photogramm. Remote Sens. 169 (2020) 268-279.

[28]

L. Dai, H. Liu, H. Tang, Z. Wu, P. Song, AO2-DETR: Arbitrary-oriented object detection transformer, IEEE Trans. Circuits Syst. Video Technol. 33 (5) (2023) 2342-2356, https://doi.org/10.1109/TCSVT.2022.3222906.

[29]

Q. Li, Y. Zhang, L. Fang, Y. Kang, S. Li, X. and Xiang Zhu, DREB-Net: Dual-stream restoration embedding blur-feature fusion network for high-mobility UAV object detection, IEEE Trans. Geosci. Remote Sens. 63 (2025) 1-18, https://doi.org/1010.1109/TCSVT.2022.3222906,Art no. 5621218.

[30]

W. Qian, X. Yang, S. Peng, et al., Learning modulated loss for rotated object detection [C], Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, 2021, pp. 2458-2466.

[31]

M. Sandler, A. Howard, M. Zhu, et al., Mobilenetv2: Inverted residuals and linear bottlenecks [C], in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 4510-4520.

[32]

A. Howard, M. Sandler, G. Chu, et al., Searching for mobilenetv3 [C], in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 1314-1324.

[33]

D. Qin, C. Leichner, M. Delakis, et al., Mobilenetv4: universal models for the mobile ecosystem [C], European Conference on Computer Vision, Springer, 2024, pp. 78-96.

[34]

M. Tan, Le Q., Efficientnet: Rethinking model scaling for convolutional neural networks [C], International Conference on Machine Learning, PMLR, 2019, pp. 6105-6114.

[35]

M. Tan, Le Q., Efficientnetv2: Smaller models and faster training [C], International Conference on Machine Learning, PMLR, 2021, pp. 10096-10106.

[36]

X. Zhang, X. Zhou, M. Lin, et al., Shufflenet: An extremely efficient convolutional neural network for mobile devices [C], in: Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 6848-6856.

[37]

N. Ma, X. Zhang, H.T. Zheng, et al., Shufflenetv2: Practical guidelines for efficient cnn architecture design[c], Proceedings of the European Conference on Computer Vision, ECCV, 2018, pp. 116-131.

[38]

K. Han, Y. Wang, Q. Tian, et al., Ghostnet: More features from cheap operations [C], in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 1580-1589.

[39]

Y. Tang, K. Han, J. Guo, et al., Ghostnetv2: Enhance cheap operation with long-range attention, Adv. Neural Inf. Process. Syst. 35 (2022) 9969-9982.

[40]

Y. Wang, Y. Wei, W. Li, et al., Research on lightweight fatigue driving detection model based on YOLOv5s, Coal Mine Mach. 46 (03) (2025) 178-181.

[41]

C. Leng, Z. Dou, H. Li, et al., Extremely low bit neural network: Squeeze the last bit out with admm [C], Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, (2018).

[42]

P. Li, Y.L. Su, H. Ning, et al., Detection of self-explosion defects in power insulators based on embedded YOLO network [J/OL], Trans. China Electrotech. Soc., 2025, pp. 1-13, https://doi.org/10.19595/J.Cn Ki.1000-6753.Tces.242110.

[43]

Zhang Ming, Li Hua, Wang Liang, DS MYOLO: A reliable object detector based on SSMs for driving scenarios [C], 2024 International Conference on Computer Vision and Pattern Recognition, CVPR. Beijing, China, 2024-06-15 To 2024-06-20, IEEE, 2024, pp. 1234-1240.

[44]

Chen Qibo, Jin Weizhong, Ge Jianyue, Liu Mengdi, Yan Yuchao, Jiang Jian, Yu Li, Guo Xuanjiang, Li Shuchang, Chen Jianzhong, CP-DETR: Concept prompt guide DETR toward stronger universal object detection [C], The Thirty-Ninth AAAI Conference on Artificial Intelligence, AAAI-25, AAAI Press, (2025).

PDF

28

Accesses

0

Citation

Detail

Sections
Recommended

/