Visual feature inter-learning for sign language recognition in emergency medicine

Chao Wei , Yunpeng Li , Jingze Liu

Optoelectronics Letters ›› 2025, Vol. 21 ›› Issue (10) : 619 -625.

PDF
Optoelectronics Letters ›› 2025, Vol. 21 ›› Issue (10) : 619 -625. DOI: 10.1007/s11801-025-4214-6
Original Paper
research-article

Visual feature inter-learning for sign language recognition in emergency medicine

Author information +
History +
PDF

Abstract

Accessible communication based on sign language recognition (SLR) is the key to emergency medical assistance for the hearing-impaired community. Balancing the capture of both local and global information in SLR for emergency medicine poses a significant challenge. To address this, we propose a novel approach based on the inter-learning of visual features between global and local information. Specifically, our method enhances the perception capabilities of the visual feature extractor by strategically leveraging the strengths of convolutional neural network (CNN), which are adept at capturing local features, and visual transformers which perform well at perceiving global features. Furthermore, to mitigate the issue of overfitting caused by the limited availability of sign language data for emergency medical applications, we introduce an enhanced short temporal module for data augmentation through additional subsequences. Experimental results on three publicly available sign language datasets demonstrate the efficacy of the proposed approach.

Keywords

A

Cite this article

Download citation ▾
Chao Wei, Yunpeng Li, Jingze Liu. Visual feature inter-learning for sign language recognition in emergency medicine. Optoelectronics Letters, 2025, 21(10): 619-625 DOI:10.1007/s11801-025-4214-6

登录浏览全文

4963

注册一个新账户 忘记密码

References

RIGHTS & PERMISSIONS

Tianjin University of Technology

AI Summary AI Mindmap
PDF

32

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/