High-resolution recognition of FOAM modes via an improved EfficientNet V2 based convolutional neural network

Youzhi Shi, Zuhai Ma, Hongyu Chen, Yougang Ke, Yu Chen, Xinxing Zhou

Front. Phys. ›› 2024, Vol. 19 ›› Issue (3) : 32205.

PDF(5192 KB)
Front. Phys. All Journals
PDF(5192 KB)
Front. Phys. ›› 2024, Vol. 19 ›› Issue (3) : 32205. DOI: 10.1007/s11467-023-1373-4
RESEARCH ARTICLE
RESEARCH ARTICLE

High-resolution recognition of FOAM modes via an improved EfficientNet V2 based convolutional neural network

Author information +
History +

Abstract

Vortex beam with fractional orbital angular momentum (FOAM) is the excellent candidate for improving the capacity of free-space optical (FSO) communication system due to its infinite modes. Therefore, the recognition of FOAM modes with higher resolution is always of great concern. In this work, through an improved EfficientNetV2 based convolutional neural network (CNN), we experimentally achieve the implementation of the recognition of FOAM modes with a resolution as high as 0.001. To the best of our knowledge, it is the first time this high resolution has been achieved. Under the strong atmospheric turbulence (AT) (Cn2=1015m2/3), the recognition accuracy of FOAM modes at 0.1 and 0.01 resolution with our model is up to 99.12% and 92.24% for a long transmission distance of 2000 m. Even for the resolution at 0.001, the recognition accuracy can still remain at 78.77%. This work provides an effective method for the recognition of FOAM modes, which may largely improve the channel capacity of the free-space optical communication.

Graphical abstract

Keywords

OAM / free-space optical communication / deep learning / convolutional neural network

Cite this article

Download citation ▾
Youzhi Shi, Zuhai Ma, Hongyu Chen, Yougang Ke, Yu Chen, Xinxing Zhou. High-resolution recognition of FOAM modes via an improved EfficientNet V2 based convolutional neural network. Front. Phys., 2024, 19(3): 32205 https://doi.org/10.1007/s11467-023-1373-4

1 Introduction

The vortex beams (VBs) can be characterized by a phase factor exp(ilφ) and the orbital angular momentum (OAM) l carried for each photon [1]. The VBs exhibit a spiral wave front and phase singularity at the center, where the l,,φ represent the topological charge, the Planck constant and the azimuthal angle respectively. Due to these special characteristics and more manipulation dimension, the VBs have highly potential applications in quantum entanglement [2], sensing [3], super-resolution microscopy [4], biomedicine [5], nonlinear optics [6], and optical communication [7]. In a free-space optical communication (FSO) system [8-10], the transmission capability can be greatly expanded with the introduction of OAM, especially the fractional orbital angular momentum (FOAM) due to the theoretically infinite number of available OAM modes [11]. However, the slight difference in the radius of the central circular ring between adjacent modes poses a great challenge to the recognition of FOAM modes compared to the integer modes [12, 13]. Furthermore, in an OAM-FSO system, the atmospheric turbulence (AT) is a common harmful factor that causes the optical intensity fluctuation, random phase perturbation, and position displacement on the VB signal [14, 15]. Therefore, the high-resolution recognition of FOAM modes under strong AT is still a problem demanding prompt solution. Various methods have been proposed to recognize integer OAM modes [16-21] and FOAM modes at 0.1 resolution [22, 23]. However, achieving higher-resolution recognition of FOAM modes remains a challenging endeavor.
In recent years, deep learning (DL) has been widely used in computer vision [24-27]. Compared to other neural networks, convolutional neural network (CNN) [28, 29] possesses superior capabilities in feature extraction and image data processing, which are the critical factors in image recognition. By building a CNN with an 8-layer structure, the recognition of FOAM modes at 0.01 resolution was firstly achieved with the recognition accuracy more than 99% [30]. However, this architecture is not suitable under strong AT because it lacks a module for extracting detailed features from images. Then, a 50-layer network structure based residual neural network (ResNet 50, a common CNN architecture) was introduced to recognize the FOAM modes at 0.1 resolution under strong AT (Cn2=5×1014m2/3) and long transmission distance (z = 1500 m). The recognition accuracy was as high as 85.30% [31]. By passing the FOAM through a column lens and then feeding it into the ResNet 50, the recognition resolution of FOAM modes was further improved to 0.01 even under strong AT (Cn2=1015m2/3) and long transmission distance (z = 1500 m). The recognition accuracy reached 99.07% at 0.1 resolution and 73.68% at 0.01 resolution [32]. However, the ResNet 50 will suffer overfitting and gradient vanishing easily when training large-scale datasets, which is unfriendly to the image recognition task with a high similarity [33]. Recently, the EfficientNet V2 based CNN has been demonstrated that it can achieve higher recognition accuracy in image classification by employing incremental learning methods and architectural improvements such as the squeeze-and-excitation (SE) module and self-attention mechanism [34-36]. Therefore, the EfficientNet V2 may help further improve the recognition accuracy of FOAM modes at higher resolution (like 0.001) even under strong AT and long transmission distance.
In this work, we experimentally achieve the recognition of FOAM modes at 0.1, 0.01, and 0.001 resolution with much high recognition accuracy under strong AT and long transmission distance by an improved EfficientNet V2 based CNN. By modifying the layers of the SE module and optimizing other parameters (dropout, width factor, depth factor, activation function, etc.), the improved model can extract more detail features of FOAM modes. Based on this proposed model, we experimentally verify its great effectiveness for the recognition of FOAM modes at 0.1 and 0.01 resolution with much high recognition accuracy. Most importantly, for the FOAM modes at 0.001 resolution, the recognition accuracy can still reach 78.77% even under strong AT (Cn2=1015m2/3) and long transmission distance (z = 2000 m). Our work exhibits the huge potential of CNN in the recognition of FOAM modes, which may promote the booming development of FSO communication.

2 Principles and concepts

2.1 Experimental setup

Fig.1 shows the experimental setup for generating and recognizing the FOAM modes. A highly stable He-Ne laser (wavelength: 632.8 nm; beam waist: 1.0 mm) is used as the Gaussian beam source. The beam is firstly expanded to 6 mm by the beam expansion system [formed by lens L1 (50 mm) and L2 (300 mm)]. A Glan−Teller (GL) prism is used to control the beam polarization. Then, the beam incidents on a spatial light modulator (SLM: Holoeye, ERIS-VIS-109), which is preloaded with the phase hologram of FOAM, AT, and AD. Finally, the diffraction pattern captured by the charge-coupled device (CCD: Thorlabs, BC106N-VIS/M) is input to the improved EfficientNet V2 model for recognizing the FOAM modes.
Fig.1 Diagram of the experimental setup. L1, L2: Lens; GL: Glan prism polarizer; SLM: Spatial light modulator; CCD: Charge-coupled device; PH: Phase hologram of FOAM; AT: Atmospheric turbulence; AD: Aperture diaphragm.

Full size|PPT slide

2.2 Intensity distribution of VBs after passing AD

FOAM modes (with topological charge l) can be easily obtained by loading the spiral wavefront phase on a Gaussian beam using SLM [37]
E(r0,θ0)=Aexp(r02w02)exp(ilθ0),
where (r0,θ0) denotes the cylindrical coordinate, A is the complex amplitude, w0 is the waist of the Gaussian beam, and l is the topological charge.
In order to obtain more detailed features of the FOAM, we design an aperture diaphragm (AD) consisting of a square hole with an internal circular screen shown in Fig.2(a). The transmittance function of AD can be expressed as
Fig.2 (a) The structure of the AD. (b) Intensity distribution of original OAM modes (l = 1−5) by simulation. (c) Intensity distribution of OAM modes (l = 1−5) diffracting after the AD by simulation. (d) Intensity distribution of original OAM modes (l = 1−5) by experiment. (e) Intensity distribution of OAM modes (l = 1−5) diffracting after the AD by experiment.

Full size|PPT slide

T(r1,θ1)=f1(r1,θ1)+f2(r1,θ1),
f1(r1,θ1)={1,r<a×sec(π4)0,others,,
f2(r1,θ1)={1,r<R,0,others,
where f1 and f2 are the transmittance functions of the square hole and circular screen, respectively, a is the side length of the square hole, and R is the radius of circular screen. In this work, R=a/2=1.5mm.
According to the Collins diffraction integral equation [38, 39], the field distribution of the VB after passing through the AD can be expressed as
U(r,θ,z)=exp(ikz)iλz002πE(r0,θ0)T(r1,θ1)×exp(ik2z[r02+r122r0r1cos(θ0θ1)])r0dr0dθ0,
where z is the transmission distance.
Finally, the beam intensity received on the CCD can be expressed as
I(r,θ,z)=U(r,θ,z)×U(r,θ,z).
To verify the effectiveness of our designed AD, we give out the simulation and experimental results of OAM modes in Fig.2. Fig.2(b) and (c) show the simulated intensity distributions of the original OAM modes (l = 1−5) and diffraction patterns after AD. In Fig.2(b), the radius of center ring increases gradually with the increase of l. After the AD, the diffraction patterns in Fig.2(c) present a four-fold symmetry [40], arising from the square diaphragm contained in AD. When l > 1, the number of diffracted spots in the outermost layer is (l + 1)/2 (when l is odd), or (l + 2)/2 (when l is even). In addition, whether there are intersection points in the center region can also determine the even or odd value of l (When l is odd, the intersection point appears). For example, when l = 3, the number of diffracted spots in the outermost layer is 2, and there are four intersection points in the center.

2.3 Construction of AT

During the beam transmission, the AT can seriously distort the wavefront phase and intensity distribution of the VB, which brings serious disturbance to the OAM communication [41, 42]. Here, we use the Hill Andrews model to simulate the AT, the equivalent phase can be shown as [43]
Φn(k)=0.33Cn2[1+1.802k2kl20.254(k2kl2)712]×exp(k2kl2)(k02+k2)116,
where k0=2π/L0, L0 is the outer scale of AT, kl=2π/l0, l0 is the inner scale of AT, k2=kx2+ky2, kx and ky denote the wavenumbers in the x and y directions, respectively, and Cn2 is the structure constant of the refractive index of air which describes the intensity of AT [44].
The spatial variation in AT (σ2(k)) can be approximated by a number of phase screens, which modify the phase profile as the beam transmission. The relationship between σ2(k) and Φn(k) can be expressed as [45]
σ2(k)=(2πNΔx)22πkz2ΔzΦn(k),
where Δx denotes the grid spacing of a random phase screen, N indicates the phase screen size, Δz is the interval distance between adjacent phase screens, and kz=2π/λ. We utilize a fast Fourier transform (FFT) operation to express the phase screen in the frequency domain,
ξ(k)=FFT(Mσ(k)),
where M is a complex Gaussian random matrix with mean 0 and variance 1 in the frequency domain, and FFT represents the fast Fourier transform operation.
During the beam transmission, the phase information of each layer can be designed and superimposed on the SLM. It should be noted that as the interval distance Δz between the adjacent phase screens decreases, the simulation accuracy increases rapidly. However, the computational resource requirement increases significantly, which is not conducive to the collection of large data sets. After several tests, we finally chose Δz= 250 m. The inner scale and outer scale of AT are set to l0 = 0.01 m and L0 = 100 m. Fig.3 shows the intensity of FOAM modes under three different AT and transmission distance (z = 2000 m) experimentally. As Cn2 increases, the intensity of FOAM modes undergoes increasingly obvious distortions which enlarges the challenge of recognizing.
Fig.3 Intensity of FOAM modes (l = 4.1, 4.3, 4.5, 4.7, 4.9) under three different intensities of AT and transmission distance (z = 2000 m). Rows (a‒c) show the intensity of the FOAM modes under AT (Cn2=1017m2/3, Cn2=1016m2/3 and Cn2=1015m2/3), respectively.

Full size|PPT slide

2.4 Architecture of the improved EfficientNet V2 based CNN

Noting that some components of the EfficientNet V2 model like SE module is useful for analyzing the detailed features of images. Therefore, for higher recognition resolution and accuracy, we increase the layers of the SE module and optimize other parameters (dropout, width factor, depth factor, activation function, etc.) to extract more detail feature of the FOAM modes.
The architecture of the CNN is shown in Fig.4(a), where the input image size is adjusted to 224 × 224, Conv3 × 3 is the 3 × 3 convolution + activation function (SiLU) + Batch Normalization (BN). The Fused-MBConv module enhances the CNN performance by combining feature maps of different resolutions. When expansion is set to 1 [Fig.4(b)], it consists of a 3 × 3 convolutional layer, a BN layer with SiLU activation, and a dropout layer with 0.2 dropout rate. In the case of expansion ≠ 1 [Fig.4(c)], it includes an additional Project Conv module. By using these two different expansion ratios (1 and ≠ 1), the CNN can achieve faster training and manage complexity efficiently in image recognition. The structure of the MBConv module is shown in Fig.4(d), compared to Fused-MBConv module, it has an additional SE module, which means it can automatically learn the importance of different channel features. 0.25 represents that the node number of the first fully connected layer in the SE module occupies 25% of the input MBConv module feature matrix channels. The key distinction between the MBConv and Fused-MBConv modules lies in the incorporation of fused point-wise and depth-wise convolution, which enhances the computational efficiency and network performance in the latter.
Fig.4 Schematic diagram of the structure of the EfficientNet V2 model. (a) Flowchart of the EfficientNet V2 model. (b) Structure diagram of the Fused-MBConv module (Expansion = 1). (c) Structure diagram of the Fused-MBConv module (Expansion ≠ 1). (d) Structure diagram of the MBConv module.

Full size|PPT slide

In the end of the CNN model, a module consisting of a 1 × 1 convolutional layer + pooling layer + FC (fully connected layer, activation = Softmax) is used as the output for the classification of the FOAM modes. As for the loss function, we use the Sparse Categorical Cross Entropy [46], which can be expressed as
loss(f(X,b),Y)=1mi=1myilog(y),
where m is the number of data set input to the CNN, y* is the predicted output, and yi is the ideal output. The weight and bias parameters are updated using the Adam optimizer with a learning rate of 0.005, aiming to minimize the loss function.

3 Results and discussion

3.1 Recognition of FOAM modes at 0.1 and 0.01 resolution

For the FOAM modes at 0.1 resolution, we experimentally collect 10000 images (80% serving as the training sets and 20% as the test sets). Fig.5(a)−(d) show the recognition results under a medium AT (Cn2=1016m2/3) and transmission distance (z = 500 m) of FOAM modes at 0.1 resolution. Fig.5(a) presents the intensity distribution of FOAM modes transmitting under AT with l = 4.1, 4.3, 4.5, 4.7, and 4.9, respectively. As l increases, the radius of central circular ring increases and the size of the notch increases first and then decreases. The corresponding intensity distributions diffracting after the ADs are shown in Fig.5(b). As l increases, the diffraction center gradually becomes a rectangular hole, and the rest of the diffracted spots gradually transition into the dart-shaped structures. It is worth noting that at this point, the topological charge l can still be recognized narrowly. After applying our proposed CNN, the recognition accuracy and loss values of the FOAM modes in the test set are shown in Fig.5(c) and (d) (red line). For comparison, we also show the situation under Cn2=1015m2/3 (blackline) and Cn2=1017m2/3 (bule line). As a result, the recognition accuracy in test set can quickly approach 100% after only 3 epochs. After 20 epochs, the accuracy and loss curves under three different AT all tend to be stable. When the epochs exceed 100, the recognition accuracy of FOAM modes finally reach 99.80%, 100%, and 100% under three different AT, while the losses are 1.25×102, 2.44×107, and 1.57×107. It is shown that our proposed CNN model exhibits perfect performance for the recognition of FOAM modes at 0.1 resolution under strong AT and long transmission distance.
Fig.5 Test results of 0.1 and 0.01 resolution FOAM modes at z = 500 m. (a, e) Original FOAM modes of 0.1 and 0.01 resolution at Cn2=1016m2/3. (b, f) Diffraction patterns detected by CCD of 0.1 and 0.01 resolution. (c, g) Accuracy curves of 0.1 and 0.01 resolution FOAM modes in test set under different AT. (d, h) Loss curves of 0.1 and 0.01 resolution FOAM modes in test set under different AT.

Full size|PPT slide

For the FOAM modes at 0.01 resolution, we experimentally collect 25000 images (80% serving as the training set and 20% as the test set), results show in Fig.5(e)‒(h). From Fig.5(e), the radius of the center ring of adjacent FOAM modes increases at a much slower rate than the FOAM modes at 0.1 resolution and the appearance of notches is barely visible. Fig.5(f) shows the intensity distribution of FOAM modes diffracting after AD. Unfortunately, it is not possible to directly distinguish the FOAM modes at 0.01 resolution. Then, our proposed CNN model is applied for the recognition of FAOM modes. From Fig.5(g) and (h), after 40 epochs, the recognition accuracy curve and loss curve in test set all become stably, the recognition accuracy of FOAM modes finally reaches 94.76%, 96.34%, and 98.46%, while the losses are 0.3411, 0.2038, and 0.0513 under three different AT. The model converges more slowly than that at 0.1 resolution. However, our model can still possess high performance at 0.01 resolution.

3.2 Recognition of FOAM modes at 0.001 resolution

Furtherly, the proposed CNN model is utilized in recognition of the FOAM modes at 0.001 resolution. Due to the more difficulty for the recognition of FAOM modes at higher resolution, we expand the number of data set to improve the recognition accuracy. We eventually collect 60 000 images, in which 48 000 are used as the training set and others as the test set. Fig.6 shows the recognition results under a medium AT of Cn2=1016m2/3 and transmission distance of z = 500 m at 0.001 resolution. Unfortunately, both from the radius of the center ring of the FOAM modes in Fig.6(a) and from the spot distribution of the FOAM modes after AD in Fig.6(b), the differences between adjacent FOAM modes are essentially invisible and can no longer be recognized. Nevertheless, our proposed CNN model can solve this problem. Fig.6(c) and (d) show the recognition accuracy and loss value in the test set under three different AT. Due to the increase in training difficulty, the recognition accuracy for the first few epochs dropped significantly comparing to the previous ones. However, since our proposed model has great advantages in performing feature extraction, it can still converge around 60 epochs. After 100 epochs, the recognition accuracy of FOAM modes under weak (Cn2=1017m2/3), medium (Cn2=1016m2/3), strong (Cn2=1015m2/3) AT are 94.88%, 91.48%, and 89.47%. The losses are 0.3102, 0.4271, and 0.4861. It can be concluded that our proposed model is still highly effective for the recognition of FOAM modes at 0.001 resolution, even under strong AT and long transmission distance.
Fig.6 Test results of 0.001 resolution FOAM modes at z = 500 m. (a) Original FOAM modes and (b) diffraction patterns detected by CCD (Cn2=1016m2/3; l = 4.001, 4.003, 4.005, 4.007, 4.009). (c) Accuracy and (d) loss curves in test set under different AT.

Full size|PPT slide

Finally, we comprehensively investigate the effects of different transmission distance and AT on the recognition accuracy of FOAM modes. Fig.7(a) shows the recognition accuracy of FOAM modes at 0.1 resolution under different AT and transmission distances. As the AT increase, the recognition accuracy decreases slightly but still remains at a very high level. Even under strong AT (Cn2=1015m2/3) and long distance (z = 2000 m), the recognition accuracy is still up to 99.12%. From the above results, our model is extremely effective in recognizing FOAM modes at 0.1 resolution. Fig.7(b) shows the recognition accuracy of FOAM modes at 0.01 resolution. Obviously, even under strong AT (Cn2=1015m2/3) and long transmission distance (z = 2000 m), the recognition accuracy reaches 92.24%. As anticipated, the recognition accuracy of FOAM modes at 0.01 resolution still remains at a high level, further confirming the great effectiveness of our model. Fig.7(c) shows the recognition accuracy of FOAM modes at 0.001 resolution. With the increase of AT and transmission distance, the recognition accuracy decreases significantly. Under the weak AT (Cn2=1017m2/3), the recognition accuracy decreases to 94.88%, 93.17%, 92.45% and 91.18% with the increase of transmission distance. Under the medium AT (Cn2=1016m2/3), the recognition accuracy decreases to 91.48%, 90.67%, 90.37% and 89.42%. Under the strong AT (Cn2=1015m2/3), the recognition accuracy decreases to 89.47%, 86.73%, 80.63% and 78.77%. The above results show that the accuracy is relate to AT and remain high value even under strong AT.
Fig.7 Test accuracy of FOAM modes for different resolution under different AT and distances. (a) 0.1 resolution. (b) 0.01 resolution. (c) 0.001 resolution.

Full size|PPT slide

In addition, compared with the ResNet 50 [32], our model shows significant improvement, especially under strong AT (Cn2=1015m2/3) and long transmission distance (z = 1500 m), the recognition accuracy is improved by 20.3% for the FOAM modes at 0.01 resolution. Furthermore, we also successfully realize the recognition of 0.001 FOAM modes. The specific comparative information for this improved EfficientNet V2 and ResNet 50 is shown in Tab.1. It can be seen that our network is superior to ResNet 50 in terms of network complexity and training time, which greatly reduces the training time and computational resources. Therefore, our model has an advantage over the ResNet 50 in handling the task of FOAM recognition.
Tab.1 Comparison of the improved EfficientNet V2 and ResNet 50 for the recognition of FOAM modes at 0.01 resolution under strong AT (Cn2=1015m2/3) and long transmission distance (z = 1500 m).
Model Params Dropout Batch size Lr Optimizer Time Accuracy
ResNet 50 25.5 M 0.2 32 0.001 Adam 78 h 73.68%
Ours 24.3 M 0.2 32 0.005 Adam 54 h 93.98%

4 Conclusion

In conclusion, we propose an improved EfficientNet V2 based CNN model for the recognition of FOAM modes experimentally. By modifying the layers of the SE module and optimizing other parameters, the model has a better ability to extract image features. By learning the mapping relationship between the topological charge l and the intensity distribution of the FOAM modes after the AD, the different FOAM modes at 0.1, 0.01 and 0.001 resolution under strong AT and long transmission distance can be precisely recognized based on the model. Specifically, for the FOAM modes at 0.1, 0.01 and 0.001 resolution, the recognition accuracy is 99.12%, 92.24% and 78.77%, respectively, even in the presence of Cn2=1015m2/3 and z = 2000 m. These findings represent significant advancements in enhancing the channel capacity of FSO communication.

References

[1]
L. Allen, M. W. Beijersbergen, R. J. C. Spreeuw, J. P. Woerdman. Orbital angular momentum of light and the transformation of Laguerre−Gaussian laser modes. Phys. Rev. A, 1992, 45(11): 8185
CrossRef ADS Google scholar
[2]
G. C. G. Berkhout, M. P. J. Lavery, J. Courtial, M. W. Beijersbergen, M. J. Padgett. Efficient sorting of orbital angular momentum states of light. Phys. Rev. Lett., 2010, 105(15): 153601
CrossRef ADS Google scholar
[3]
K. Liu, Y. Q. Cheng, X. Li, Y. Gao. Microwave-sensing technology using orbital angular momentum: Overview of its advantages. IEEE Veh. Technol. Mag., 2019, 14(2): 112
CrossRef ADS Google scholar
[4]
L. Yan, P. Kristensen, S. Ramachandran. Vortex fibers for STED microscopy. APL Photonics, 2019, 4(2): 022903
CrossRef ADS Google scholar
[5]
X. W. Zhuang. Unraveling DNA condensation with optical tweezers. Science, 2004, 305(5681): 188
CrossRef ADS Google scholar
[6]
Z. Y. Zhou, D. S. Ding, Y. K. Jiang, Y. Li, S. Shi, X. S. Wang, B. S. Shi. Orbital angular momentum light frequency conversion and interference with quasi-phase matching crystals. Opt. Express, 2014, 22(17): 20298
CrossRef ADS Google scholar
[7]
S. J. Li, Z. Y. Li, G. S. Huang, X. B. Liu, R. Q. Li, X. Y. Cao. Digital coding transmissive metasurface for multi-OAM-beam. Front. Phys., 2022, 17(6): 62501
CrossRef ADS Google scholar
[8]
L. Zou, L. Wang, S. M. Zhao. Turbulence mitigation scheme based on spatial diversity in orbital-angular-momentum multiplexed system. Opt. Commun., 2017, 400: 123
CrossRef ADS Google scholar
[9]
E.M. AmhoudM. ChafiiA.NimrG.Fettweis, OFDM with index modulation in orbital angular momentum multiplexed free space optical links, in: IEEE 93rd Vehicular Technology Conference (VTC-Spring), Electr Network, 2021
[10]
A. E. Willner, K. Pang, H. Song, K. H. Zou, H. B. Zhou. Orbital angular momentum of light for communications. Appl. Phys. Rev., 2021, 8(4): 041312
CrossRef ADS Google scholar
[11]
X. H. Zhang, T. Xia, S. B. Cheng, S. H. Tao. Free-space information transfer using the elliptic vortex beam with fractional topological charge. Opt. Commun., 2019, 431: 238
CrossRef ADS Google scholar
[12]
V. V. Kotlyar, A. A. Kovalev, A. G. Nalimov, A. P. Porfirev. Evolution of an optical vortex with an initial fractional topological charge. Phys. Rev. A, 2020, 102(2): 023516
CrossRef ADS Google scholar
[13]
S. S. Li, B. F. Shen, W. P. Wang, Z. G. Bu, H. Zhang, H. Zhang, S. H. Zhai. Diffraction of relativistic vortex harmonics with fractional average orbital angular momentum. Chin. Opt. Lett., 2019, 17(5): 050501
CrossRef ADS Google scholar
[14]
M. I. Dedo, Z. Wang, K. Guo, Y. Sun, F. Shen, H. Zhou, J. Gao, R. Sun, Z. Ding, Z. Guo. Retrieving performances of vortex beams with GS algorithm after transmitting in different types of turbulences. Appl. Sci. (Basel), 2019, 9(11): 2269
CrossRef ADS Google scholar
[15]
X. Yan, P. F. Zhang, J. H. Zhang, X. X. Feng, C. H. Qiao, C. Y. Fan. Effect of atmospheric turbulence on entangled orbital angular momentum three-qubit state. Chin. Phys. B, 2017, 26(6): 064202
CrossRef ADS Google scholar
[16]
Y. J. Yang, Q. Zhao, L. L. Liu, Y. D. Liu, C. Rosales-Guzman, C. W. Qiu. Manipulation of orbital-angular-momentum spectrum using pinhole plates. Phys. Rev. Appl., 2019, 12(6): 064007
CrossRef ADS Google scholar
[17]
Z. C. Zhang, J. C. Pei, Y. P. Wang, X. G. Wang. Measuring orbital angular momentum of vortex beams in optomechanics. Front. Phys., 2021, 16(3): 32503
CrossRef ADS Google scholar
[18]
A. Forbes, A. Dudley, M. McLaren. Creation and detection of optical modes with spatial light modulators. Adv. Opt. Photonics, 2016, 8(2): 200
CrossRef ADS Google scholar
[19]
J.YuZ.F. Wang, 3D facial motion tracking by combining online appearance model and cylinder head model in particle filtering, Sci. China Inf. Sci. 57(7), 029101 (2014)
[20]
N. Uribe-Patarroyo, A. Fraine, D. S. Simon, O. Minaeva, A. V. Sergienko. Object identification using correlated orbital angular momentum states. Phys. Rev. Lett., 2013, 110(4): 043601
CrossRef ADS Google scholar
[21]
J. Zhu, P. Zhang, D. Z. Fu, D. X. Chen, R. F. Liu, Y. N. Zhou, H. Gao, F. L. Li. Probing the fractional topological charge of a vortex light beam by using dynamic angular double slits. Photon. Res., 2016, 4(5): 187
CrossRef ADS Google scholar
[22]
D. Deng, M. C. Lin, Y. Li, H. Zhao. Precision measurement of fractional orbital angular momentum. Phys. Rev. Appl., 2019, 12(1): 014048
CrossRef ADS Google scholar
[23]
S. Zheng, J. Wang. Measuring orbital angular momentum (OAM) states of vortex beams with annular gratings. Sci. Rep., 2017, 7(1): 40781
CrossRef ADS Google scholar
[24]
K.BayoudhR. KnaniF.HamdaouiA.Mtibaa, A survey on deep multimodal learning for computer vision: Advances, trends, applications, and datasets, Vis. Comput. 38(8), 2939 (2022)
[25]
N.O’MahonyS.CampbellA.Carvalho S.HarapanahalliG.V. HernandezL.KrpalkovaD.RiordanJ.Walsh, Deep learning vs. traditional computer vision, in: Computer Vision Conference (CVC), Springer International Publishing Ag, Las Vegas, NV, 2019, pp 128–144
[26]
J. Long, E. Shelhamer, and T. Darrell, Fully convolutional networks for semantic segmentation, in: IEEE Conference on Computer, Vision and Pattern Recognition (CVPR), IEEE, Boston, MA, 2015, pp 3431–3440
[27]
N. Le, V. S. Rathour, K. Yamazaki, K. Luu, M. Savvides. Deep reinforcement learning in computer vision: a comprehensive survey. Artif. Intell. Rev., 2022, 55(4): 2733
CrossRef ADS Google scholar
[28]
R. Yamashita, M. Nishio, R. K. G. Do, K. Togashi. Convolutional neural networks: An overview and application in radiology. Insights Imaging, 2018, 9(4): 611
CrossRef ADS Google scholar
[29]
P.MichalskiB. RuszczakM.Tomaszewski, Convolutional neural networks implementations for computer vision, in: 3rd International Scientific Conference on Brain-Computer Interfaces (BCI), Springer International Publishing Ag, Opole Univ Technol, Opole, POLAND, 2018, pp 98–110
[30]
Z. W. Liu, S. Yan, H. G. Liu, X. F. Chen. Superhigh-resolution recognition of optical vortex modes assisted by a deep-learning method. Phys. Rev. Lett., 2019, 123(18): 183902
CrossRef ADS Google scholar
[31]
M. Cao, Y. L. Yin, J. W. Zhou, J. H. Tang, L. P. Cao, Y. Xia, J. P. Yin. Machine learning based accurate recognition of fractional optical vortex modes in atmospheric environment. Appl. Phys. Lett., 2021, 119(14): 141103
CrossRef ADS Google scholar
[32]
J. Zhou, Y. Yin, J. Tang, C. Ling, M. Cao, L. Cao, G. Liu, J. Yin, Y. Xia. Recognition of high-resolution optical vortex modes with deep residual learning. Phys. Rev. A, 2022, 106(1): 013519
CrossRef ADS Google scholar
[33]
W. W. Song, S. T. Li, L. Y. Fang, T. Lu. Hyperspectral image classification with deep feature fusion network. IEEE Trans. Geosci. Remote Sens., 2018, 56(6): 3173
CrossRef ADS Google scholar
[34]
M.X. TanQ. V. Le, EfficientNetV2: Smaller models and faster training, in: International Conference on Machine Learning (ICML), Electr Network, 2021, pp 7102–7110
[35]
M.L. HuangY. C. Liao, A lightweight CNN-based network on COVID-19 detection using X-ray and CT images, Comput. Biol. Med. 146, 105604 (2022)
[36]
R. Karthik, T. S. Vaichole, S. K. Kulkarni, O. Yadav, F. Khan. Eff2Net: An efficient channel attention-based convolutional neural network for skin disease classification. Biomed. Signal Process. Control, 2022, 73: 103406
CrossRef ADS Google scholar
[37]
H. Zhang, J. Zeng, X. Y. Lu, Z. Y. Wang, C. L. Zhao, Y. J. Cai. Review on fractional vortex beam. Nanophotonics, 2022, 11(2): 241
CrossRef ADS Google scholar
[38]
A. Belafhal, L. Dalil-Essakali. Collins formula and propagation of Bessel-modulated Gaussian light beams through an ABCD optical system. Opt. Commun., 2000, 177(1−6): 181
CrossRef ADS Google scholar
[39]
Y. J. Yang, Y. Dong, C. L. Zhao, Y. J. Cai. Generation and propagation of an anomalous vortex beam. Opt. Lett., 2013, 38(24): 5418
CrossRef ADS Google scholar
[40]
P. H. F. Mesquita, A. J. Jesus-Silva, E. J. S. Fonseca, J. M. Hickmann. Engineering a square truncated lattice with light’s orbital angular momentum. Opt. Express, 2011, 19(21): 20616
CrossRef ADS Google scholar
[41]
B. Rodenburg, M. P. J. Lavery, M. Malik, M. N. O’Sullivan, M. Mirhosseini, D. J. Robertson, M. Padgett, R. W. Boyd. Influence of atmospheric turbulence on states of light carrying orbital angular momentum. Opt. Lett., 2012, 37(17): 3735
CrossRef ADS Google scholar
[42]
S. Y. Fu, C. Q. Gao. Influences of atmospheric turbulence effects on the orbital angular momentum spectra of vortex beams. Photon. Res., 2016, 4(5): B1
CrossRef ADS Google scholar
[43]
L. C. Andrews. An analytical model for the refractive index power spectrum and its application to optical scintillations in the atmosphere. J. Mod. Opt., 1992, 39(9): 1849
CrossRef ADS Google scholar
[44]
W. Cheng, J. W. Haus, Q. W. Zhan. Propagation of vector vortex beams through a turbulent atmosphere. Opt. Express, 2009, 17(20): 17829
CrossRef ADS Google scholar
[45]
S. M. Zhao, J. Leach, L. Y. Gong, J. Ding, B. Y. Zheng. Aberration corrections for free-space optical communications in atmosphere turbulence using orbital angular momentum states. Opt. Express, 2012, 20(1): 452
CrossRef ADS Google scholar
[46]
Y. Kim, I. Ohn, D. Kim. Fast convergence rates of deep neural networks for classification. Neural Netw., 2021, 138: 179
CrossRef ADS Google scholar

Declarations

The authors declare that they have no competing interests and there are no conflicts.

Acknowledgements

This work was supported by the National Natural Science Foundation of China (Grant Nos. 62271332, 12374273, and 62275162), the Guangdong Basic and Applied Basic Research Foundation (No. 2023A1515030152), the Shenzhen Government’s Plan of Science and Technology (Nos. JCYJ20180305124927623 and JCYJ20190808150205481), and the Training Program for Excellent Young innovators of Changsha (No. kq2107013).

RIGHTS & PERMISSIONS

2024 Higher Education Press
AI Summary AI Mindmap
PDF(5192 KB)

695

Accesses

4

Citations

Detail

Sections
Recommended

/