High-resolution recognition of FOAM modes via an improved EfficientNet V2 based convolutional neural network

Youzhi Shi , Zuhai Ma , Hongyu Chen , Yougang Ke , Yu Chen , Xinxing Zhou

Front. Phys. ›› 2024, Vol. 19 ›› Issue (3) : 32205

PDF (5192KB)
Front. Phys. ›› 2024, Vol. 19 ›› Issue (3) : 32205 DOI: 10.1007/s11467-023-1373-4
RESEARCH ARTICLE

High-resolution recognition of FOAM modes via an improved EfficientNet V2 based convolutional neural network

Author information +
History +
PDF (5192KB)

Abstract

Vortex beam with fractional orbital angular momentum (FOAM) is the excellent candidate for improving the capacity of free-space optical (FSO) communication system due to its infinite modes. Therefore, the recognition of FOAM modes with higher resolution is always of great concern. In this work, through an improved EfficientNetV2 based convolutional neural network (CNN), we experimentally achieve the implementation of the recognition of FOAM modes with a resolution as high as 0.001. To the best of our knowledge, it is the first time this high resolution has been achieved. Under the strong atmospheric turbulence (AT) ( Cn2= 1015m2/3), the recognition accuracy of FOAM modes at 0.1 and 0.01 resolution with our model is up to 99.12% and 92.24% for a long transmission distance of 2000 m. Even for the resolution at 0.001, the recognition accuracy can still remain at 78.77%. This work provides an effective method for the recognition of FOAM modes, which may largely improve the channel capacity of the free-space optical communication.

Graphical abstract

Keywords

OAM / free-space optical communication / deep learning / convolutional neural network

Cite this article

Download citation ▾
Youzhi Shi, Zuhai Ma, Hongyu Chen, Yougang Ke, Yu Chen, Xinxing Zhou. High-resolution recognition of FOAM modes via an improved EfficientNet V2 based convolutional neural network. Front. Phys., 2024, 19(3): 32205 DOI:10.1007/s11467-023-1373-4

登录浏览全文

4963

注册一个新账户 忘记密码

1 Introduction

The vortex beams (VBs) can be characterized by a phase factor e xp(i lφ) and the orbital angular momentum (OAM) l carried for each photon [1]. The VBs exhibit a spiral wave front and phase singularity at the center, where the l,,φ represent the topological charge, the Planck constant and the azimuthal angle respectively. Due to these special characteristics and more manipulation dimension, the VBs have highly potential applications in quantum entanglement [2], sensing [3], super-resolution microscopy [4], biomedicine [5], nonlinear optics [6], and optical communication [7]. In a free-space optical communication (FSO) system [8-10], the transmission capability can be greatly expanded with the introduction of OAM, especially the fractional orbital angular momentum (FOAM) due to the theoretically infinite number of available OAM modes [11]. However, the slight difference in the radius of the central circular ring between adjacent modes poses a great challenge to the recognition of FOAM modes compared to the integer modes [12, 13]. Furthermore, in an OAM-FSO system, the atmospheric turbulence (AT) is a common harmful factor that causes the optical intensity fluctuation, random phase perturbation, and position displacement on the VB signal [14, 15]. Therefore, the high-resolution recognition of FOAM modes under strong AT is still a problem demanding prompt solution. Various methods have been proposed to recognize integer OAM modes [16-21] and FOAM modes at 0.1 resolution [22, 23]. However, achieving higher-resolution recognition of FOAM modes remains a challenging endeavor.

In recent years, deep learning (DL) has been widely used in computer vision [24-27]. Compared to other neural networks, convolutional neural network (CNN) [28, 29] possesses superior capabilities in feature extraction and image data processing, which are the critical factors in image recognition. By building a CNN with an 8-layer structure, the recognition of FOAM modes at 0.01 resolution was firstly achieved with the recognition accuracy more than 99% [30]. However, this architecture is not suitable under strong AT because it lacks a module for extracting detailed features from images. Then, a 50-layer network structure based residual neural network (ResNet 50, a common CNN architecture) was introduced to recognize the FOAM modes at 0.1 resolution under strong AT ( C n2= 5× 10 14 m2/3) and long transmission distance (z = 1500 m). The recognition accuracy was as high as 85.30% [31]. By passing the FOAM through a column lens and then feeding it into the ResNet 50, the recognition resolution of FOAM modes was further improved to 0.01 even under strong AT ( Cn2=1015 m2/3) and long transmission distance (z = 1500 m). The recognition accuracy reached 99.07% at 0.1 resolution and 73.68% at 0.01 resolution [32]. However, the ResNet 50 will suffer overfitting and gradient vanishing easily when training large-scale datasets, which is unfriendly to the image recognition task with a high similarity [33]. Recently, the EfficientNet V2 based CNN has been demonstrated that it can achieve higher recognition accuracy in image classification by employing incremental learning methods and architectural improvements such as the squeeze-and-excitation (SE) module and self-attention mechanism [34-36]. Therefore, the EfficientNet V2 may help further improve the recognition accuracy of FOAM modes at higher resolution (like 0.001) even under strong AT and long transmission distance.

In this work, we experimentally achieve the recognition of FOAM modes at 0.1, 0.01, and 0.001 resolution with much high recognition accuracy under strong AT and long transmission distance by an improved EfficientNet V2 based CNN. By modifying the layers of the SE module and optimizing other parameters (dropout, width factor, depth factor, activation function, etc.), the improved model can extract more detail features of FOAM modes. Based on this proposed model, we experimentally verify its great effectiveness for the recognition of FOAM modes at 0.1 and 0.01 resolution with much high recognition accuracy. Most importantly, for the FOAM modes at 0.001 resolution, the recognition accuracy can still reach 78.77% even under strong AT (Cn2=1015m2/3) and long transmission distance (z = 2000 m). Our work exhibits the huge potential of CNN in the recognition of FOAM modes, which may promote the booming development of FSO communication.

2 Principles and concepts

2.1 Experimental setup

Fig.1 shows the experimental setup for generating and recognizing the FOAM modes. A highly stable He-Ne laser (wavelength: 632.8 nm; beam waist: 1.0 mm) is used as the Gaussian beam source. The beam is firstly expanded to 6 mm by the beam expansion system [formed by lens L1 (50 mm) and L2 (300 mm)]. A Glan−Teller (GL) prism is used to control the beam polarization. Then, the beam incidents on a spatial light modulator (SLM: Holoeye, ERIS-VIS-109), which is preloaded with the phase hologram of FOAM, AT, and AD. Finally, the diffraction pattern captured by the charge-coupled device (CCD: Thorlabs, BC106N-VIS/M) is input to the improved EfficientNet V2 model for recognizing the FOAM modes.

2.2 Intensity distribution of VBs after passing AD

FOAM modes (with topological charge l) can be easily obtained by loading the spiral wavefront phase on a Gaussian beam using SLM [37]

E(r0, θ 0)=Aexp( r02 w02)exp( il θ0),

where (r0, θ0) denotes the cylindrical coordinate, A is the complex amplitude, w0 is the waist of the Gaussian beam, and l is the topological charge.

In order to obtain more detailed features of the FOAM, we design an aperture diaphragm (AD) consisting of a square hole with an internal circular screen shown in Fig.2(a). The transmittance function of AD can be expressed as

T(r1, θ 1)=f1(r1,θ1)+ f2(r1,θ1),

f1(r1,θ1)= { 1,r< a× s ec(π4)0, o th er s,,

f2(r1,θ1)= { 1,r< R,0 ,others,

where f1 and f2 are the transmittance functions of the square hole and circular screen, respectively, a is the side length of the square hole, and R is the radius of circular screen. In this work, R=a/2=1.5m m.

According to the Collins diffraction integral equation [38, 39], the field distribution of the VB after passing through the AD can be expressed as

U(r, θ,z)= exp( ikz)iλ z002πE( r0, θ 0)T( r1, θ 1) × ex p(ik 2z [r02+ r122r0r1cos( θ0 θ1)]) r0dr0dθ0,

where z is the transmission distance.

Finally, the beam intensity received on the CCD can be expressed as

I(r, θ,z)=U( r,θ,z)× U (r,θ ,z).

To verify the effectiveness of our designed AD, we give out the simulation and experimental results of OAM modes in Fig.2. Fig.2(b) and (c) show the simulated intensity distributions of the original OAM modes (l = 1−5) and diffraction patterns after AD. In Fig.2(b), the radius of center ring increases gradually with the increase of l. After the AD, the diffraction patterns in Fig.2(c) present a four-fold symmetry [40], arising from the square diaphragm contained in AD. When l > 1, the number of diffracted spots in the outermost layer is (l + 1)/2 (when l is odd), or (l + 2)/2 (when l is even). In addition, whether there are intersection points in the center region can also determine the even or odd value of l (When l is odd, the intersection point appears). For example, when l = 3, the number of diffracted spots in the outermost layer is 2, and there are four intersection points in the center.

2.3 Construction of AT

During the beam transmission, the AT can seriously distort the wavefront phase and intensity distribution of the VB, which brings serious disturbance to the OAM communication [41, 42]. Here, we use the Hill Andrews model to simulate the AT, the equivalent phase can be shown as [43]

Φn(k)= 0.33 Cn2 [1+1.802 k2kl20.254(k2 kl2 ) 7 12]×exp( k 2 k l2)( k02+ k2)116,

where k0= 2π/L0, L0 is the outer scale of AT, kl=2π/l0, l0 is the inner scale of AT, k2=kx2+ ky2, kx and ky denote the wavenumbers in the x and y directions, respectively, and Cn2 is the structure constant of the refractive index of air which describes the intensity of AT [44].

The spatial variation in AT (σ2(k) ) can be approximated by a number of phase screens, which modify the phase profile as the beam transmission. The relationship between σ2(k) and Φn(k) can be expressed as [45]

σ 2 (k)= (2 πNΔx)22 πkz2Δz Φn(k),

where Δx denotes the grid spacing of a random phase screen, N indicates the phase screen size, Δz is the interval distance between adjacent phase screens, and kz=2π/λ. We utilize a fast Fourier transform (FFT) operation to express the phase screen in the frequency domain,

ξ(k)=FFT (Mσ (k)),

where M is a complex Gaussian random matrix with mean 0 and variance 1 in the frequency domain, and FFT represents the fast Fourier transform operation.

During the beam transmission, the phase information of each layer can be designed and superimposed on the SLM. It should be noted that as the interval distance Δz between the adjacent phase screens decreases, the simulation accuracy increases rapidly. However, the computational resource requirement increases significantly, which is not conducive to the collection of large data sets. After several tests, we finally chose Δz= 250 m. The inner scale and outer scale of AT are set to l0 = 0.01 m and L0 = 100 m. Fig.3 shows the intensity of FOAM modes under three different AT and transmission distance (z = 2000 m) experimentally. As Cn2 increases, the intensity of FOAM modes undergoes increasingly obvious distortions which enlarges the challenge of recognizing.

2.4 Architecture of the improved EfficientNet V2 based CNN

Noting that some components of the EfficientNet V2 model like SE module is useful for analyzing the detailed features of images. Therefore, for higher recognition resolution and accuracy, we increase the layers of the SE module and optimize other parameters (dropout, width factor, depth factor, activation function, etc.) to extract more detail feature of the FOAM modes.

The architecture of the CNN is shown in Fig.4(a), where the input image size is adjusted to 224 × 224, Conv3 × 3 is the 3 × 3 convolution + activation function (SiLU) + Batch Normalization (BN). The Fused-MBConv module enhances the CNN performance by combining feature maps of different resolutions. When expansion is set to 1 [Fig.4(b)], it consists of a 3 × 3 convolutional layer, a BN layer with SiLU activation, and a dropout layer with 0.2 dropout rate. In the case of expansion ≠ 1 [Fig.4(c)], it includes an additional Project Conv module. By using these two different expansion ratios (1 and ≠ 1), the CNN can achieve faster training and manage complexity efficiently in image recognition. The structure of the MBConv module is shown in Fig.4(d), compared to Fused-MBConv module, it has an additional SE module, which means it can automatically learn the importance of different channel features. 0.25 represents that the node number of the first fully connected layer in the SE module occupies 25% of the input MBConv module feature matrix channels. The key distinction between the MBConv and Fused-MBConv modules lies in the incorporation of fused point-wise and depth-wise convolution, which enhances the computational efficiency and network performance in the latter.

In the end of the CNN model, a module consisting of a 1 × 1 convolutional layer + pooling layer + FC (fully connected layer, activation = Softmax) is used as the output for the classification of the FOAM modes. As for the loss function, we use the Sparse Categorical Cross Entropy [46], which can be expressed as

loss(f(X,b) ,Y)=1m i=1m yil og ( y),

where m is the number of data set input to the CNN, y* is the predicted output, and yi is the ideal output. The weight and bias parameters are updated using the Adam optimizer with a learning rate of 0.005, aiming to minimize the loss function.

3 Results and discussion

3.1 Recognition of FOAM modes at 0.1 and 0.01 resolution

For the FOAM modes at 0.1 resolution, we experimentally collect 10000 images (80% serving as the training sets and 20% as the test sets). Fig.5(a)−(d) show the recognition results under a medium AT (Cn2=1016 m2/3) and transmission distance (z = 500 m) of FOAM modes at 0.1 resolution. Fig.5(a) presents the intensity distribution of FOAM modes transmitting under AT with l = 4.1, 4.3, 4.5, 4.7, and 4.9, respectively. As l increases, the radius of central circular ring increases and the size of the notch increases first and then decreases. The corresponding intensity distributions diffracting after the ADs are shown in Fig.5(b). As l increases, the diffraction center gradually becomes a rectangular hole, and the rest of the diffracted spots gradually transition into the dart-shaped structures. It is worth noting that at this point, the topological charge l can still be recognized narrowly. After applying our proposed CNN, the recognition accuracy and loss values of the FOAM modes in the test set are shown in Fig.5(c) and (d) (red line). For comparison, we also show the situation under Cn2=10 15 m2/3 (blackline) and Cn2=10 17 m2/3 (bule line). As a result, the recognition accuracy in test set can quickly approach 100% after only 3 epochs. After 20 epochs, the accuracy and loss curves under three different AT all tend to be stable. When the epochs exceed 100, the recognition accuracy of FOAM modes finally reach 99.80%, 100%, and 100% under three different AT, while the losses are 1.25×102, 2.44×107, and 1.57×107. It is shown that our proposed CNN model exhibits perfect performance for the recognition of FOAM modes at 0.1 resolution under strong AT and long transmission distance.

For the FOAM modes at 0.01 resolution, we experimentally collect 25000 images (80% serving as the training set and 20% as the test set), results show in Fig.5(e)‒(h). From Fig.5(e), the radius of the center ring of adjacent FOAM modes increases at a much slower rate than the FOAM modes at 0.1 resolution and the appearance of notches is barely visible. Fig.5(f) shows the intensity distribution of FOAM modes diffracting after AD. Unfortunately, it is not possible to directly distinguish the FOAM modes at 0.01 resolution. Then, our proposed CNN model is applied for the recognition of FAOM modes. From Fig.5(g) and (h), after 40 epochs, the recognition accuracy curve and loss curve in test set all become stably, the recognition accuracy of FOAM modes finally reaches 94.76%, 96.34%, and 98.46%, while the losses are 0.3411, 0.2038, and 0.0513 under three different AT. The model converges more slowly than that at 0.1 resolution. However, our model can still possess high performance at 0.01 resolution.

3.2 Recognition of FOAM modes at 0.001 resolution

Furtherly, the proposed CNN model is utilized in recognition of the FOAM modes at 0.001 resolution. Due to the more difficulty for the recognition of FAOM modes at higher resolution, we expand the number of data set to improve the recognition accuracy. We eventually collect 60 000 images, in which 48 000 are used as the training set and others as the test set. Fig.6 shows the recognition results under a medium AT of Cn2= 1016m2/3 and transmission distance of z = 500 m at 0.001 resolution. Unfortunately, both from the radius of the center ring of the FOAM modes in Fig.6(a) and from the spot distribution of the FOAM modes after AD in Fig.6(b), the differences between adjacent FOAM modes are essentially invisible and can no longer be recognized. Nevertheless, our proposed CNN model can solve this problem. Fig.6(c) and (d) show the recognition accuracy and loss value in the test set under three different AT. Due to the increase in training difficulty, the recognition accuracy for the first few epochs dropped significantly comparing to the previous ones. However, since our proposed model has great advantages in performing feature extraction, it can still converge around 60 epochs. After 100 epochs, the recognition accuracy of FOAM modes under weak (Cn2= 1017m2/3), medium (Cn2= 1016m2/3), strong (Cn2= 1015m2/3) AT are 94.88%, 91.48%, and 89.47%. The losses are 0.3102, 0.4271, and 0.4861. It can be concluded that our proposed model is still highly effective for the recognition of FOAM modes at 0.001 resolution, even under strong AT and long transmission distance.

Finally, we comprehensively investigate the effects of different transmission distance and AT on the recognition accuracy of FOAM modes. Fig.7(a) shows the recognition accuracy of FOAM modes at 0.1 resolution under different AT and transmission distances. As the AT increase, the recognition accuracy decreases slightly but still remains at a very high level. Even under strong AT (Cn2= 1015m2/3) and long distance (z = 2000 m), the recognition accuracy is still up to 99.12%. From the above results, our model is extremely effective in recognizing FOAM modes at 0.1 resolution. Fig.7(b) shows the recognition accuracy of FOAM modes at 0.01 resolution. Obviously, even under strong AT (Cn2=1015 m2/3) and long transmission distance (z = 2000 m), the recognition accuracy reaches 92.24%. As anticipated, the recognition accuracy of FOAM modes at 0.01 resolution still remains at a high level, further confirming the great effectiveness of our model. Fig.7(c) shows the recognition accuracy of FOAM modes at 0.001 resolution. With the increase of AT and transmission distance, the recognition accuracy decreases significantly. Under the weak AT (Cn2= 1017m2/3), the recognition accuracy decreases to 94.88%, 93.17%, 92.45% and 91.18% with the increase of transmission distance. Under the medium AT (Cn2= 1016m2/3), the recognition accuracy decreases to 91.48%, 90.67%, 90.37% and 89.42%. Under the strong AT (Cn2= 1015m2/3), the recognition accuracy decreases to 89.47%, 86.73%, 80.63% and 78.77%. The above results show that the accuracy is relate to AT and remain high value even under strong AT.

In addition, compared with the ResNet 50 [32], our model shows significant improvement, especially under strong AT (Cn2=10 15 m2/3) and long transmission distance (z = 1500 m), the recognition accuracy is improved by 20.3% for the FOAM modes at 0.01 resolution. Furthermore, we also successfully realize the recognition of 0.001 FOAM modes. The specific comparative information for this improved EfficientNet V2 and ResNet 50 is shown in Tab.1. It can be seen that our network is superior to ResNet 50 in terms of network complexity and training time, which greatly reduces the training time and computational resources. Therefore, our model has an advantage over the ResNet 50 in handling the task of FOAM recognition.

4 Conclusion

In conclusion, we propose an improved EfficientNet V2 based CNN model for the recognition of FOAM modes experimentally. By modifying the layers of the SE module and optimizing other parameters, the model has a better ability to extract image features. By learning the mapping relationship between the topological charge l and the intensity distribution of the FOAM modes after the AD, the different FOAM modes at 0.1, 0.01 and 0.001 resolution under strong AT and long transmission distance can be precisely recognized based on the model. Specifically, for the FOAM modes at 0.1, 0.01 and 0.001 resolution, the recognition accuracy is 99.12%, 92.24% and 78.77%, respectively, even in the presence of Cn2= 1015m2/3 and z = 2000 m. These findings represent significant advancements in enhancing the channel capacity of FSO communication.

References

[1]

L. Allen, M. W. Beijersbergen, R. J. C. Spreeuw, J. P. Woerdman. Orbital angular momentum of light and the transformation of Laguerre−Gaussian laser modes. Phys. Rev. A, 1992, 45(11): 8185

[2]

G. C. G. Berkhout, M. P. J. Lavery, J. Courtial, M. W. Beijersbergen, M. J. Padgett. Efficient sorting of orbital angular momentum states of light. Phys. Rev. Lett., 2010, 105(15): 153601

[3]

K. Liu, Y. Q. Cheng, X. Li, Y. Gao. Microwave-sensing technology using orbital angular momentum: Overview of its advantages. IEEE Veh. Technol. Mag., 2019, 14(2): 112

[4]

L. Yan, P. Kristensen, S. Ramachandran. Vortex fibers for STED microscopy. APL Photonics, 2019, 4(2): 022903

[5]

X. W. Zhuang. Unraveling DNA condensation with optical tweezers. Science, 2004, 305(5681): 188

[6]

Z. Y. Zhou, D. S. Ding, Y. K. Jiang, Y. Li, S. Shi, X. S. Wang, B. S. Shi. Orbital angular momentum light frequency conversion and interference with quasi-phase matching crystals. Opt. Express, 2014, 22(17): 20298

[7]

S. J. Li, Z. Y. Li, G. S. Huang, X. B. Liu, R. Q. Li, X. Y. Cao. Digital coding transmissive metasurface for multi-OAM-beam. Front. Phys., 2022, 17(6): 62501

[8]

L. Zou, L. Wang, S. M. Zhao. Turbulence mitigation scheme based on spatial diversity in orbital-angular-momentum multiplexed system. Opt. Commun., 2017, 400: 123

[9]

E.M. AmhoudM. ChafiiA.NimrG.Fettweis, OFDM with index modulation in orbital angular momentum multiplexed free space optical links, in: IEEE 93rd Vehicular Technology Conference (VTC-Spring), Electr Network, 2021

[10]

A. E. Willner, K. Pang, H. Song, K. H. Zou, H. B. Zhou. Orbital angular momentum of light for communications. Appl. Phys. Rev., 2021, 8(4): 041312

[11]

X. H. Zhang, T. Xia, S. B. Cheng, S. H. Tao. Free-space information transfer using the elliptic vortex beam with fractional topological charge. Opt. Commun., 2019, 431: 238

[12]

V. V. Kotlyar, A. A. Kovalev, A. G. Nalimov, A. P. Porfirev. Evolution of an optical vortex with an initial fractional topological charge. Phys. Rev. A, 2020, 102(2): 023516

[13]

S. S. Li, B. F. Shen, W. P. Wang, Z. G. Bu, H. Zhang, H. Zhang, S. H. Zhai. Diffraction of relativistic vortex harmonics with fractional average orbital angular momentum. Chin. Opt. Lett., 2019, 17(5): 050501

[14]

M. I. Dedo, Z. Wang, K. Guo, Y. Sun, F. Shen, H. Zhou, J. Gao, R. Sun, Z. Ding, Z. Guo. Retrieving performances of vortex beams with GS algorithm after transmitting in different types of turbulences. Appl. Sci. (Basel), 2019, 9(11): 2269

[15]

X. Yan, P. F. Zhang, J. H. Zhang, X. X. Feng, C. H. Qiao, C. Y. Fan. Effect of atmospheric turbulence on entangled orbital angular momentum three-qubit state. Chin. Phys. B, 2017, 26(6): 064202

[16]

Y. J. Yang, Q. Zhao, L. L. Liu, Y. D. Liu, C. Rosales-Guzman, C. W. Qiu. Manipulation of orbital-angular-momentum spectrum using pinhole plates. Phys. Rev. Appl., 2019, 12(6): 064007

[17]

Z. C. Zhang, J. C. Pei, Y. P. Wang, X. G. Wang. Measuring orbital angular momentum of vortex beams in optomechanics. Front. Phys., 2021, 16(3): 32503

[18]

A. Forbes, A. Dudley, M. McLaren. Creation and detection of optical modes with spatial light modulators. Adv. Opt. Photonics, 2016, 8(2): 200

[19]

J.YuZ.F. Wang, 3D facial motion tracking by combining online appearance model and cylinder head model in particle filtering, Sci. China Inf. Sci. 57(7), 029101 (2014)

[20]

N. Uribe-Patarroyo, A. Fraine, D. S. Simon, O. Minaeva, A. V. Sergienko. Object identification using correlated orbital angular momentum states. Phys. Rev. Lett., 2013, 110(4): 043601

[21]

J. Zhu, P. Zhang, D. Z. Fu, D. X. Chen, R. F. Liu, Y. N. Zhou, H. Gao, F. L. Li. Probing the fractional topological charge of a vortex light beam by using dynamic angular double slits. Photon. Res., 2016, 4(5): 187

[22]

D. Deng, M. C. Lin, Y. Li, H. Zhao. Precision measurement of fractional orbital angular momentum. Phys. Rev. Appl., 2019, 12(1): 014048

[23]

S. Zheng, J. Wang. Measuring orbital angular momentum (OAM) states of vortex beams with annular gratings. Sci. Rep., 2017, 7(1): 40781

[24]

K.BayoudhR. KnaniF.HamdaouiA.Mtibaa, A survey on deep multimodal learning for computer vision: Advances, trends, applications, and datasets, Vis. Comput. 38(8), 2939 (2022)

[25]

N.O’MahonyS.CampbellA.Carvalho S.HarapanahalliG.V. HernandezL.KrpalkovaD.RiordanJ.Walsh, Deep learning vs. traditional computer vision, in: Computer Vision Conference (CVC), Springer International Publishing Ag, Las Vegas, NV, 2019, pp 128–144

[26]

J. Long, E. Shelhamer, and T. Darrell, Fully convolutional networks for semantic segmentation, in: IEEE Conference on Computer, Vision and Pattern Recognition (CVPR), IEEE, Boston, MA, 2015, pp 3431–3440

[27]

N. Le, V. S. Rathour, K. Yamazaki, K. Luu, M. Savvides. Deep reinforcement learning in computer vision: a comprehensive survey. Artif. Intell. Rev., 2022, 55(4): 2733

[28]

R. Yamashita, M. Nishio, R. K. G. Do, K. Togashi. Convolutional neural networks: An overview and application in radiology. Insights Imaging, 2018, 9(4): 611

[29]

P.MichalskiB. RuszczakM.Tomaszewski, Convolutional neural networks implementations for computer vision, in: 3rd International Scientific Conference on Brain-Computer Interfaces (BCI), Springer International Publishing Ag, Opole Univ Technol, Opole, POLAND, 2018, pp 98–110

[30]

Z. W. Liu, S. Yan, H. G. Liu, X. F. Chen. Superhigh-resolution recognition of optical vortex modes assisted by a deep-learning method. Phys. Rev. Lett., 2019, 123(18): 183902

[31]

M. Cao, Y. L. Yin, J. W. Zhou, J. H. Tang, L. P. Cao, Y. Xia, J. P. Yin. Machine learning based accurate recognition of fractional optical vortex modes in atmospheric environment. Appl. Phys. Lett., 2021, 119(14): 141103

[32]

J. Zhou, Y. Yin, J. Tang, C. Ling, M. Cao, L. Cao, G. Liu, J. Yin, Y. Xia. Recognition of high-resolution optical vortex modes with deep residual learning. Phys. Rev. A, 2022, 106(1): 013519

[33]

W. W. Song, S. T. Li, L. Y. Fang, T. Lu. Hyperspectral image classification with deep feature fusion network. IEEE Trans. Geosci. Remote Sens., 2018, 56(6): 3173

[34]

M.X. TanQ. V. Le, EfficientNetV2: Smaller models and faster training, in: International Conference on Machine Learning (ICML), Electr Network, 2021, pp 7102–7110

[35]

M.L. HuangY. C. Liao, A lightweight CNN-based network on COVID-19 detection using X-ray and CT images, Comput. Biol. Med. 146, 105604 (2022)

[36]

R. Karthik, T. S. Vaichole, S. K. Kulkarni, O. Yadav, F. Khan. Eff2Net: An efficient channel attention-based convolutional neural network for skin disease classification. Biomed. Signal Process. Control, 2022, 73: 103406

[37]

H. Zhang, J. Zeng, X. Y. Lu, Z. Y. Wang, C. L. Zhao, Y. J. Cai. Review on fractional vortex beam. Nanophotonics, 2022, 11(2): 241

[38]

A. Belafhal, L. Dalil-Essakali. Collins formula and propagation of Bessel-modulated Gaussian light beams through an ABCD optical system. Opt. Commun., 2000, 177(1−6): 181

[39]

Y. J. Yang, Y. Dong, C. L. Zhao, Y. J. Cai. Generation and propagation of an anomalous vortex beam. Opt. Lett., 2013, 38(24): 5418

[40]

P. H. F. Mesquita, A. J. Jesus-Silva, E. J. S. Fonseca, J. M. Hickmann. Engineering a square truncated lattice with light’s orbital angular momentum. Opt. Express, 2011, 19(21): 20616

[41]

B. Rodenburg, M. P. J. Lavery, M. Malik, M. N. O’Sullivan, M. Mirhosseini, D. J. Robertson, M. Padgett, R. W. Boyd. Influence of atmospheric turbulence on states of light carrying orbital angular momentum. Opt. Lett., 2012, 37(17): 3735

[42]

S. Y. Fu, C. Q. Gao. Influences of atmospheric turbulence effects on the orbital angular momentum spectra of vortex beams. Photon. Res., 2016, 4(5): B1

[43]

L. C. Andrews. An analytical model for the refractive index power spectrum and its application to optical scintillations in the atmosphere. J. Mod. Opt., 1992, 39(9): 1849

[44]

W. Cheng, J. W. Haus, Q. W. Zhan. Propagation of vector vortex beams through a turbulent atmosphere. Opt. Express, 2009, 17(20): 17829

[45]

S. M. Zhao, J. Leach, L. Y. Gong, J. Ding, B. Y. Zheng. Aberration corrections for free-space optical communications in atmosphere turbulence using orbital angular momentum states. Opt. Express, 2012, 20(1): 452

[46]

Y. Kim, I. Ohn, D. Kim. Fast convergence rates of deep neural networks for classification. Neural Netw., 2021, 138: 179

RIGHTS & PERMISSIONS

Higher Education Press

AI Summary AI Mindmap
PDF (5192KB)

1211

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/