Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
CFANet: Context Feature Fusion and Attention Mechanism Based Network for Small Target Segmentation in Medical Images
Previous Article in Journal
A Comprehensive Analysis: Evaluating Security Characteristics of Xbee Devices against Zigbee Protocol
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

An Efficient Recognition Method for Orbital Angular Momentum via Adaptive Deep ELM

School of Computer Science and Technology, Changchun University of Science and Technology, Changchun 130022, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8737; https://doi.org/10.3390/s23218737
Submission received: 11 September 2023 / Revised: 18 October 2023 / Accepted: 24 October 2023 / Published: 26 October 2023
(This article belongs to the Section Physical Sensors)

Abstract

:
For orbital angular momentum (OAM) recognition in atmosphere turbulence, how to design a self-adapted model is a challenging problem. To address this issue, an efficient deep learning framework that uses a derived extreme learning machine (ELM) has been put forward. Different from typical neural network methods, the provided analytical machine learning model can match the different OAM modes automatically. In the model selection phase, a multilayer ELM is adopted to quantify the laser spot characteristics. In the parameter optimization phase, a fast iterative shrinkage-thresholding algorithm makes the model present the analytic expression. After the feature extraction of the received intensity distributions, the proposed method develops a relationship between laser spot and OAM mode, thus building the steady neural network architecture for the new received vortex beam. The whole recognition process avoids the trial and error caused by user intervention, which makes the model suitable for a time-varying atmospheric environment. Numerical simulations are conducted on different experimental datasets. The results demonstrate that the proposed method has a better capacity for OAM recognition.

1. Introduction

Optical carrier technology has received great attention for high-rate information transmission [1]. The orbital angular momentum (OAM) provides abundant carrier information, which is widely used in Free Space Optical (FSO) communication systems [2]. However, as an important factor in FSO, atmosphere turbulence causes the distortion of optical signals. It will lead to the propagation phenomena such as laser spot scintillation, laser beam drift, and arrival angle fluctuation [3]. These negative effects increase the difficulty of optical signal processing. If the received laser spot information cannot be extracted effectively, the transmission capacity and spectral efficiency will be limited [4]. Therefore, how to recognize the OAM modes in different atmosphere turbulence environment has been regarded as a challenging problem.
Because the vortex beam with different superposition states presents various physical phenomena, the OAM mode recognition in turbulences can extend the channel capacity of communication systems. Taking advantage of the data-driven strategy, many researchers devote themselves to developing the OAM mode detection schemes. For example, Krenn et al. [5] firstly use the Artificial Neural Network (ANN) as an OAM mode classification. Then, Krenn’s team conducts four OAM recognition experiments in a maritime atmospheric channel and confirms that machine learning can support the information transmission over hundreds of kilometers [6]. Sun et al. use support vector machine (SVM) to classify the OAM of a Laguerre–Gaussian beam [7]. In order to express the subtle features, deep learning is usually used as a modeling tool for unstructured data [8]. Jing et al. propose a fractional OAM mode recognition method with a neural network frame, and extend the original optical features [9]. Meanwhile, Fu et al. propose a hybrid interference–convolutional neural network scheme [10]. They use a deep learning framework to analyze atmosphere turbulence and improve recognition performance under different turbulence levels. Driven by miscellaneous applications, Zhao et al. propose an OAM detection strategy based on a diffractive multilayered neural network [11]. Furthermore, Zhou et al. retrain a convolutional neural network (CNN) to explore hidden OAM information in big data environments [12]. A spatial diversity turbulence mitigation scheme is designed to maintain the performance of an OAM-multiplexed communication link [13]. It is obvious that the machine learning approach can achieve the OAM mode classification effectively. However, the complex iteration process of parameters makes it difficult for the neural network to obtain an analytical solution. It causes trouble to build the appropriate model for different dynamic turbulence scenes.
Recently, as a typical feed-forward neural network, extreme learning machine (ELM) has significant advantages including the theoretical analytical solution and fast convergence speed, thus improving the feasibility of image processing [14]. In the training stage, ELM calculates the output weight by Moore–Penrose (M-P) generalized inverse, which avoids the complicated parameter ergodic process [15]. In a similar approach to [15], Huang et al. [16] use a penalty function to balance model accuracy and complexity. This penalty regression ELM overcomes the defect that the least square method is not suitable for nonlinear fitting. Tavakoli et al. attempt to develop a novel fixation prediction framework based on inter-image similarities [17]. The ELM model estimates the saliency of the given image and proves the effect of image classification. At an atmospheric turbulent environment, the recognition model of OAM mode will be appropriate for the changing scenario. To acquire the expected generalization performance, the error minimized ELM (EM-ELM) moves toward the predictive target with varying hidden node [18]. Then, D-ELM has a huge advantage on data fitting when the target function is not explicit, and the compact input set reduces the model size [19]. A recent multilayer perceptron method [20] shows a significant performance improvement for complex feature mapping. In a word, think of the intricate relationship between laser spot and OAM mode, where the updated mechanism of recognition model is a vital condition to guarantee the global optimum.
Different from the above previous work, in this paper, we present an efficient OAM recognition method with adaptive deep ELM (AD-ELM). Expanding the original ELM, the proposed method not only describes the subtle feature, but also fully considers the relationship between OAM and laser spot. Also, in order to accelerate the calculation processing, we bring in the fast iterative shrinkage-thresholding algorithm (FISTA) [21] to obtain convergence results.
This paper contains several contributions as follows:
(1)
Different from the previous job, this paper not only considers the randomness of parameters in the neural network as a whole [22], but also obtains the minimum norm constraint of the global solution.
(2)
Then, the updated model will be self-driven, that means OAM mode recognition will obtain the analytical expression. The whole learning process avoided manual parameter tuning.
(3)
It has significant application value for OAM mode recognition according to the atmospheric turbulent environment.
The whole thesis has four main parts. Part 1 is the introduction. Part 2 involves atmosphere turbulence and ELM. Part 3 interprets the proposed AD-ELM method. The last part is the results.

2. Preliminaries

The proposed method takes root in the atmosphere turbulence theory and ELM model. The atmosphere turbulence theory is the key condition to guarantee the performance and efficiency of wireless optical communication systems. The ELM has a complete theoretical system and easy hardware implementation, and it becomes feasible for unstructured data processing and image recognition [23]. In this section, the two parts are outlined.

2.1. Atmosphere Turbulence Theory

The atmospheric refractive index is one of the most efficient means to represent atmospheric conditions [24]. The structure constant of the refractive index C n 2 helps us to describe atmosphere turbulence. Based on the Kolmogorov theory [25], for the isotropic turbulence, the refraction index D n ( r ) can be described by the C n 2 , which is shown as follows:
D n ( r ) = [ n ( r + r 1 ) n ( r 1 ) ] 2 = C n 2 r 2 / 3 , l 0 < r < L 0 ,
where r 1 denotes the initial position, n denotes the refractive index, l 0 is the internal scale, and L 0 is the external scale. The symbol represents the average value. For example, the Hufnagel–Valley model expresses the daytime atmosphere turbulence. In this model, the atmospheric refractive index and wind speed are adjustable. They fit various site conditions to match specific values for the coherence length and the isoplanatic angle [26]. Assuming that the laser is traveling forward along the z axis, the atmospheric coherent parameter r 0 can be expressed as [27]
r 0 = [ 0.423 k 2 0 Δ z C n 2 ( z ) d z ] 5 / 3 ,
where Δ z is the thickness of each thin layer. We can find that the parameter r 0 describes the phase disturbance intensity. The smaller r 0 is, the stronger the disturbance becomes.

2.2. Brief of ELM Model

According to the ELM construction, the weights between the input layer and hidden layer are determined randomly, and the output vector is calculated analytically [28]. Suppose that the sample data { ( x i , y i ) } i = 1 N , the mapping relationship can be expressed as
y i = p = 1 L β p φ ( a p x i + b p ) ,   i = 1 , , N ,
where L is the number of hidden nodes, a p is the input weight vector, and b p is the offset. β p is the learning parameter connecting the pth hidden node to the output node. Normally, in a more concise expression, the above expression can be written in a matrix form as follows:
Y = H β ,
where H = [ φ ( a 1 x 1 + b 1 ) φ ( a L x 1 + b L ) φ ( a 1 x N + b 1 ) φ ( a L x N + b L ) ] N × L , β = [ β 1 , β 2 , , β L ] Τ and Y = [ y 1 , y 2 , , y N ] Τ . Since there is no iteration operation at solution procedure, the ELM is faster than the traditional gradient descent method.

3. Preliminaries

In this section, we introduce the proposed AD-ELM in detail. This method includes two phases. One of them is the model selection, and the other one is the parameter estimation. With the basic ELM, the deep learning is selected to establish a suitable OAM recognition model. We implemented numerical simulation and received light spots through a phase screen. The schematic diagram of the method is illustrated in Figure 1.

3.1. Model Selection Phase

Model selection is necessary to obtain the OAM analytic expression by means of the data-driven theory. We have simulated the Laguerre–Gaussian (LG) beams, which propagated in the random atmosphere turbulence channel. The complex amplitude expression is shown as follows:
U l ( r , φ , z ) = R l ( r , z ) exp ( i l φ ) ,
where i is the imaginary unit, and l is the topological charge. Otherwise, l is known as spatial modes. The number of l is the OAM mode label. exp ( i l φ ) is the phase factor. The radial basis function R l ( r , z ) is then given by
R l ( r , z ) = 2 p ! π ( p + | l | ) ! 1 ω ( z ) [ r 2 ω ( z ) ] | l | L p l [ 2 r 2 ω 2 ( z ) ] exp [ r 2 ω 2 ( z ) ] × exp [ i k r 2 z 2 ( z 2 + z R 2 ) ] exp [ i ( 2 p + | l | + 1 ) arctan ( z z R ) ] ,
where p is the radial index, ω ( z ) = ω 0 1 + ( z 2 / z R 2 ) , ω 0 is the waist radius, z R = π ω 0 2 / λ is the Rayleigh distance, λ is the wavelength, and L p l ( ) is the Laguerre polynomial. Considering the radial index p = 0 and topological charge number ± l 0 , the intensity of light field can be further expressed according to the Helmholtz equation
| U ± l 0 ( r , z ) | 2 = | U l 0 ( r , φ , z ) + U l 0 ( r , φ , z ) | 2 = | R l 0 ( r , z ) exp ( i l 0 φ ) + R l 0 ( r , z ) exp ( i l 0 φ ) | 2 = 2 | R | l 0 | ( r , z ) | 2 ( 1 + cos ( 2 l 0 φ ) ) ,
where the | R | l 0 | ( r , z ) | 2 can be expressed as
| R | l 0 | ( r , z ) | 2 = 2 π | l 0 | ! 1 ω 2 ( z ) exp [ 2 ( r 2 ω 2 ( z ) ) ] [ 2 r ω ( z ) ] 2 | l 0 | .
Usually, the proposed method generates the random phase screen based on fast Fourier transform (FFT) and inverse fast Fourier transform (IFFT) [29], which is shown as follows:
U ( r 1 ) = F 1 [ F [ U ( r 0 ) ] exp [ G ] ] exp [ i ϕ ( r 0 ) ] U ( r 2 ) = F 1 [ F [ U ( r 1 ) ] exp [ G ] ] exp [ i ϕ ( r 1 ) ] U ( r m ) = F 1 [ F [ U ( r m 1 ) ] exp [ G ] ] exp [ i ϕ ( r m 1 ) ] ,
where U ( r 0 ) is the intensity of the light field at the source, U ( r 1 ) is the intensity of the light field when the beam has passed through a phase screen, and U ( r m ) represents the light field intensity of the received plane. F and F 1 express the Fourier transform and the inverse Fourier transform, G = i Δ z ( 2 g 2 g r 2 ) / ( 2 g ) is transformation operator, Δ z is the distance of phase screen at each layer, g r 2 is the number of spatial waves, and r 0 = ( x 0 , y 0 ) , r 1 = ( x 1 , y 1 ) , , r m = ( x m , y m ) denote coordinates. Here, the grid size of the laser spot is 256 × 256 . C n 2 = r 0 ( 5 / 3 ) / 0.423 k 2 is equal to the structure constant of the refractive index. k = 2 π / λ is the number of laser beams. λ is the wavelength as above. After N times Monte Carlo simulation, the sample set of laser spots is P M , N = { U i ( r m ) | i = 1 , 2 , N , M = m × m } . Then, the laser spot P M , N can be converted to feature vectors P M , N = [ p M , 1 , p M , 2 , , p M , N ] . The OAM mode label is Y N = [ y 1 , y 2 , , y N ] Τ . The training set can be described as [ P M , N , Y N ] = { ( p M , i , y i ) } i = 1 N .
After receiving the sample set of laser spots, the relationship between the feature vectors and OAM mode label can be calculated by multilayer ELM. The expression can be defined analytically:
Y N = H β = H 1 β 1 β 2 β J ,
where β = ( β 1 , β 2 , β j , β J ) is the output weight, and H = ( H 1 , H 2 , H j , H J ) is the hidden output. H 1 shows the polynomial structure constituted by weight a 1 = [ a 1 , 1 , a 1 , 2 , , a 1 , L 1 ] T and offset b 1 = [ b 1 , 1 , b 1 , 2 , , b 1 , L 1 ] T , which can be expressed as
H 1 = h ( P M , N ) = [ φ ( a 1 , 1 p M , 1 + b 1 , 1 ) φ ( a 1 , L 1 p M , 1 + b 1 , L 1 ) φ ( a 1 , 1 p M , N + b 1 , 1 ) φ ( a 1 , L 1 p M , N + b 1 , L 1 ) ] N × L 1 .
The model contains multiple hidden layers. Meanwhile, each of them has independent weights. Multilayer mapping extracts feature sets automatically. Thus, the output of each hidden layer is determined by the previous layer value, which can be expressed as
H j = H j 1 β j 1 ,
where H j and H j 1 represent the output matrix for the jth layer and ( j 1 ) th layer.
The current feature becomes the input element of the next layer. When the previous layer confirms mapping relationship, the current vector is determined as well. Then, the weight expression of the output layer will be obtained. In order to improve the generalization ability, we added norm constraints to each hidden layer. The optimization object function is transformed into the form minimization between the approximation error and regularization, which can be expressed as
ε = F ( β ) = arg   min β { C H β Y N 2 + β 1 } ,
where C is the regularization parameter. It can balance the approximation error and model adaptability.

3.2. Parameter Estimation Phase

In order to achieve a parameter estimation of the output layer, the Fast Iterative Shrinkage-Thresholding Algorithm (FISTA) [21] is selected to accelerate the calculation. Taking the whole hidden layers into consideration, the J th output vectors will be calculated, and the approximation error part can be abstracted to g ( β )
g ( β ) = H β Y N 2 ,
where β = ( β 1 , β 2 , β j , β J ) is output weight. The auxiliary parameter s is the output weight of the previous layer, which means s j = β j 1 R n . After determined the interval of each iteration, the relationship between β j and s j becomes
β j = P L ( s j ) ,
where P L ( ) specific expression is described according to the minimum lower bound of the derivative. When the gradient of approximation function satisfies the Lipschitz condition, there exists a constant L P satisfying the following formula:
g ( β ) g ( s ) 2 2 L P β s 2 2 , β , s R n ,
where represents the differential operator. Using Taylor formula expansion and gradient descent method, we obtain auxiliary sequence s j + 1 = s j 1 L P f ( s j ) . It helps calculate the optimal solution of F ( β ) . The quadratic minimization form is shown as follows:
P L ( s ) = arg   min β { g ( s ) + β s , g ( s ) + L P 2 β s 2 2 + β 1 , β , s R n } .
After ignoring the irrelevant constant term, the simplified optimization equation can be obtained as
P L ( s ) = arg   min β { L P 2 β ( s 1 L P f ( s ) ) 2 + β 1 } .
During the iteration of each layer, the appropriate step size θ j of each layer is calculated from the gradient, with f being a smooth convex function. θ 1 = 1 is the initial point. For each output layer of the neural network, the above equation can be converted to the following formula:
P L ( s j ) = arg   min β { 1 2 θ j β ( s j θ j f ( s j ) ) 2 + β l 1 } .
In order to maintain the convergence, the appropriate step size θ j can be calculated recursively from the parameter values of the previous layer:
θ j = 1 + 1 + 4 θ j 1 2 2 .
The update expression of the auxiliary parameter s j can be expressed as
β j = P L ( s j + 1 ) = P L ( β j + ( θ j 1 θ j + 1 ) ( β j β j 1 ) )
After the iterative steps, the parameters of multilayer ELM were finally determined. Considering the characteristics of the learning procedure, the whole AD-ELM algorithm is presented in Algorithm 1.
Algorithm 1: Learning Procedure of the AD-ELM
1. Generate Laguerre Gaussian beams
2. Calculate the Phase Screen at objective position
3. Building the mapping relationship between laser spots and OAM mode
4. Establish multilayer ELM network structure
5. Solve the output weight of each layer
6. Return the analytic solution of the output weight
7. Identify OAM in the new simple set

4. Simulation Results and Discussion

In this section, the simulation results have been described in detail. The dataset consists of four simulation conditions. Then, the proposed AD-ELM is compared with four other classical OAM recognition methodologies.

4.1. Dataset Generation

To validate the effectiveness of the proposed method, the continuous phase screen is chosen to describe the atmospheric condition. We simulate the 2000m transmission distance. The visualized laser spot diagram is displayed in Figure 2.
Figure 2a–d shows the original laser spot and received laser spot for single OAM and multiplexed OAM. Davis proposed a method to divide the strength of turbulence according to C n 2 [30]. When the C n 2 < 6.4 × 10 17   m 2 / 3 , the atmospheric turbulence is a weak turbulence. As C n 2 increases, the atmospheric turbulence gradually becomes stronger. In order to evaluate the effectiveness of the proposed method, we generate different datasets in which C n 2 are 1 × 10 14   m 2 / 3 and 1 × 10 15   m 2 / 3 , respectively. After image processing, the received laser spot images retain the key characteristic information, and the spot size of each attribute is 28 × 28 . We simulated 2500 OAM modal samples. Here, the 2000 training samples and 500 test samples are randomly collected as the input of the model. For the single topological charges, the ten OAM modes that we use in our simulation process are given by | l 0 | = 1 , 2 , 10 . And for the multiplexed topological charges, the eight OAM modes that we use in our simulation are given by | l 0 | = 1 , 2 , 8 . The intensity images of the received LG beam in 1 × 10 14   m 2 / 3 are shown in Figure 3.
Then, all the features must be normalized into the range [−1,1]. The detailed formula is shown as follows:
x = 2 × x x min x max x min 1 ,
where x is an original numeric value, x min is the minimum of the current character data, and x max is the maximum of the current character data. Once the training process is completed, the model will perform a recognition operation at the test period.

4.2. Evaluation Criteria

The generalization performance for various methods is analyzed with the Root Mean Square Error (RMSE) and G-mean. The RMSE can be computed from the following criteria.
RMSE = 1 N t i = 1 N t ( y i y ^ i ) 2 .
where y i is the ith actual value, i = 1 , , N t . y ^ i is the calculated value by the model, N t is the number of testing records. The G-mean is calculated by the confusion matrix in Table 1.
For the whole dataset with M t labels, they are shown as the following forms:
G m e a n = ( i = 1 M t T P i T P i + F N i ) 1 M t
The higher G-mean, the better recognition performance for multi-classification data.

4.3. Simulation Results

The whole forecasting process is executed by 20 Monte Carlo trials. The BP-ANN, SVM, kNN and DNN are chosen to be the contrastive methods. We firstly analyzed the recognition result for different OAM modes. Then, the visualized classification distribution result of the proposed method is displayed in Figure 4.
We found the various categories in Figure 4. The different color spots with various color circles show the different datasets. The pink circle, gray circle, blue circle, dark blue circle, light blue circle, dark green circle, green circle, light green circle, purple circle and orange circle indicate the correct identification from the first mode to the tenth mode. Triangles of the corresponding color indicate false identification. We first calculated the principal components, and then sorted them from the largest to the smallest. After calculating the eigenvectors for each sample separately, the coordinates in two directions represent the two largest principal component scores, respectively. The proposed method distinguishes the different species. Further discussion of the results, the RMSE and G-mean of the proposed method and the comparative methods are shown as the following figure. Figure 5 shows the average RMSE for the two turbulence scenarios as various color lines. We use blue bars to represent the proposed method, and the red ones, the yellow ones, the green ones, and the pink ones represent BPP-ANN, SVM, kNN, and DNN, respectively.
In Figure 5, we can find that the proposed method has the minimum RMSE, which achieves 0.1732 and 0.0894 for C n 2 = 1 × 10 14   m 2 / 3 and C n 2 = 1 × 10 15   m 2 / 3 environment, respectively. It is proved that the proposed AD-ELM has better adaptive capacity. This advantage will describe the model analytically based on the global optimal parameters. To reflect the predicted multiple OAM recognition performance, we calculated the G-mean of the five methods, which is shown in Figure 6.
From Figure 6a,b, it is clear that the G-mean of the proposed method is always at the top position, which achieves 0.9585 and 0.9915. Especially in Figure 6a, the G-mean of the proposed method is close to 1, which is far higher than the constructive methods. The kNN shows good performance in a weak turbulence environment. If the turbulence becomes strong, the performance will degrade dramatically. It should be noted that due to the controlling weight parameters along with the turbulence variation, the proposed method predicts a high-quality annular shape effectively. What is more, the performances of recognizing multiplexed OAM mode sets are further discussed. Figure 7 shows the visualized classification distribution of the proposed method for eight OAM modes.
Following the simulation setup shown in Figure 7, we can find that the proposed method distinguishes the different categories effectively. The different color spots with various color circles show the different datasets. The pink circle, gray circle, blue circle, dark blue circle, light blue circle, dark green circle, green circle, light green circle, purple circle and orange circle indicate the correct identification from the first mode to the tenth mode. Triangles of the corresponding color indicate false identification. The sample presents dispersed distribution in strong turbulent environment. When the turbulence becomes weak, the distinguishable spatial distribution appears. Then, Figure 8 shows the RMSE of the five methods for multiplexed topological charges dataset.
Analysis of RMSE sees that the proposed method has smaller generalization errors in the whole testing process, which are 0.3847 and 0.0632. If the C n 2 becomes higher, the changeable weight will match the nonlinear feature effectively. Because of the quantitative analytical expression between the adjacent layers, the parameters will be self-adapted and the performance fluctuation is not violent. Then, we measured the G-mean for multiplexed topological charges dataset, which is shown in Figure 9.
From Figure 9, we can find that the G-mean of proposed method is 0.7849 and 0.9979. It considered different turbulent environments. As seen in Figure 9b, the contrastive methods show the good value in the weak turbulence environment. When the turbulence turns strong, the proposed AD-ELM also reaches the optimal model to match the changing environment. At the same time, a better analytical expression is given by the AD-ELM, which represents a higher model capacity without manual trial and error operation, which have significant application values for long-term online recognition. The optimal parameters help our method be more suitable for the regular OAM mode sets.
From the simulation result, the proposed method receives better generalization performance. It is because holding the optimal parameters acquires the optimal analytic solution. So, the OAM recognition will be supported by the adaptive model.

5. Conclusions

In this paper, an efficient AD-ELM is proposed for orbital angular momentum recognition. Different from a traditional neural network, the proposed method quantified the laser spot characteristics. We brought a fast iterative shrinkage-thresholding into the iterative solution procedure. We receive integrated weight automatically. So, the whole process of OAM recognition requires no manual trial and error. Compared with several contrasting methods, the simulation result shows the proposed method has better recognition results. It indicated the method matches the new facula sample well. The OAM with high-dimensional feature space will be researched in the future.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s23218737/s1.

Author Contributions

Conceptualization, H.Y. (Haiyang Yu) and C.C.; Methodology, H.Y. (Haiyang Yu); Software, C.C.; Formal analysis, C.C.; Data curation, X.H.; Writing—original draft, H.Y. (Haiyang Yu); Writing—review & editing, H.Y. (Huamin Yang). All authors have read and agreed to the published version of the manuscript.

Funding

The work received the support of the National Natural Science Foundation of China under Grant 62305030 and U19A2063, the Development Program of Science and Technology of Jilin Province of China YDZJ202101ZYTS151. The authors gratefully acknowledge financial support from the Key Laboratory of Optical Control and Optical Information Transmission Technology, Ministry of Education.

Data Availability Statement

The data presented in this study are available in Supplementary Material.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khalighi, M.A.; Uysal, M. Survey on free space optical communication: A communication theory perspective. IEEE Commun. Surv. Tutor. 2014, 16, 2231–2258. [Google Scholar] [CrossRef]
  2. Chen, R.; Zhou, H.; Moretti, M.; Wang, X.; Li, J. Orbital Angular Momentum Waves: Generation, Detection and Emerging Applications. IEEE Commun. Surv. Tutor. 2019, 22, 840–868. [Google Scholar] [CrossRef]
  3. Chen, C.; Yang, H.; Tong, S.; Lou, Y. Two-frequency mutual coherence function for Gaussian-beam pulses propagating along a horizontal path in weak anisotropic atmosphere turbulence. Appl. Opt. 2015, 54, 5797–5804. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, C. Multi-Sector Oriented Object Detector for Accurate Localization in Optical Remote Sensing Images. Remote Sens. 2021, 13, 1921. [Google Scholar]
  5. Krenn, M.; Fickler, R.; Fink, M.; Handsteiner, J.; Malik, M.; Scheidl, T.; Ursin, R.; Zeilinger, A. Communication with spatially modulated light through turbulent air across Vienna. New J. Phys. 2014, 16, 113028. [Google Scholar] [CrossRef]
  6. Krenn, M.; Handsteiner, J.; Fink, M.; Fickler, R.; Ursin, R.; Malik, M.; Zeilinger, A. Twisted light transmission over 143 km. Proc. Natl. Acad. Sci. USA 2016, 113, 13648–13653. [Google Scholar] [CrossRef]
  7. Sun, R.; Guo, L.; Cheng, M.; Li, J.; Yan, X. Identifying orbital angular momentum modes in turbulence with high accuracy via machine learning. J. Opt. 2019, 21, 075703. [Google Scholar] [CrossRef]
  8. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef]
  9. Jing, G.; Chen, L.; Wang, P.; Xiong, W.; Huang, Z.; Liu, J.; Chen, Y.; Li, Y.; Fan, D.; Chen, S. Recognizing Fractional Orbital Angular Momentum Using Feedforward Neural Network. Results Phys. 2021, 28, 104619. [Google Scholar] [CrossRef]
  10. Fu, X.; Bai, Y.H.; Yang, Y.J. Measuring OAM by the hybrid scheme of interference and convolutional neural network. Opt. Eng. 2021, 60, 064109. [Google Scholar] [CrossRef]
  11. Krenn, M.; Handsteiner, J.; Fink, M.; Fickler, R.; Ursin, R.; Malik, M.; Zeilinger, A. Orbital angular momentum detection based on diffractive deep neural network. Opt. Commun. 2019, 443, 245–249. [Google Scholar]
  12. Zhou, H.; Wang, Y.; Li, X.; Xu, Z.; Li, X.; Huang, L. A deep learning approach for trustworthy high-fidelity computational holographic orbital angular momentum communication. Appl. Phys. Lett. 2021, 119, 044104. [Google Scholar] [CrossRef]
  13. Zou, L.; Wang, L.; Zhao, S. Turbulence mitigation scheme based on spatial diversity in orbital-angular-momentum multiplexed system. Opt. Commun. 2017, 400, 123–127. [Google Scholar] [CrossRef]
  14. Huang, G.; Zhu, Q.; Siew, C. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  15. Wang, X.; Han, M. Improved extreme learning machine for multivariate time series online sequential prediction. Eng. Appl. Artif. Intell. 2015, 40, 28–36. [Google Scholar] [CrossRef]
  16. Huang, G.-B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2011, 42, 513–529. [Google Scholar] [CrossRef] [PubMed]
  17. Tavakoli, H.R.; Borji, A.; Laaksonen, J.; Rahtu, E. Exploiting inter-image similarity and ensemble of extreme learners for fixation prediction using deep features. Neurocomputing 2017, 244, 10–18. [Google Scholar] [CrossRef]
  18. Feng, G.; Huang, G.-B.; Lin, Q.; Gay, R. Error Minimized Extreme Learning Machine with Growth of Hidden Nodes and Incremental Learning. IEEE Trans. Neural Netw. 2009, 20, 1352–1357. [Google Scholar] [CrossRef]
  19. Zhang, R.; Lan, Y.; Huang, G.-B.; Xu, Z.-B.; Soh, Y.C. Dynamic extreme learning machine and its approximation capability. IEEE Trans. Cybern. 2013, 43, 2054–2065. [Google Scholar] [CrossRef]
  20. Liu, Z.; Yang, S.; Feng, Z.; Gao, Q.; Wang, M. Fast SAR Autofocus Based on Ensemble Convolutional Extreme Learning Machine. Remote Sens. 2021, 13, 2683. [Google Scholar] [CrossRef]
  21. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
  22. Yu, H.Y.; Chen, C.Y.; Yang, H.M. Two-Stage Game Strategy for Multiclass Imbalanced Data Online Prediction. Neural Process. Lett. 2020, 52, 2493–2512. [Google Scholar] [CrossRef]
  23. Bai, Z.; Huang, G.-B.; Wang, D.; Wang, H.; Westover, M.B. Sparse Extreme Learning Machine for Classification. IEEE Trans. Cybern. 2014, 44, 1858–1870. [Google Scholar] [CrossRef] [PubMed]
  24. Basu, S. A simple approach for estimating the refractive index structure parameter (Cn2) profile in the atmosphere. Opt. Lett. 2015, 40, 4130–4133. [Google Scholar] [CrossRef] [PubMed]
  25. Delgado-Vences, F.J.; Flandoli, F. A spectral-based numerical method for Kolmogorov equations in Hilbert spaces. Infin. Dimens. Anal. Quantum Probab. Relat. Top. 2016, 19, 1–37. [Google Scholar] [CrossRef]
  26. Fayed, H.A.; Abd El Aziz Aly, A.M.; Aly, M.H. Irradiance scintillation index on slant atmosphere turbulence: Simple approach. Opt. Eng. 2016, 55, 056113. [Google Scholar] [CrossRef]
  27. Peng, J.; Zhang, L.; Zhang, K.; Ma, J. Channel capacity of OAM based FSO communication systems with partially coherent Bessel–Gaussian beams in anisotropic turbulence. Opt. Commun. 2018, 418, 32–36. [Google Scholar] [CrossRef]
  28. Savitha, R.; Suresh, S.; Kim, H.J. A Meta-Cognitive Learning Algorithm for an Extreme Learning Machine Classifier. Cogn. Comput. 2014, 6, 253–263. [Google Scholar] [CrossRef]
  29. Fatheddin, P.; Gustafsson, J. Generation of a sequence of correlated phase screens. Opt. Commun. 2017, 391, 100–105. [Google Scholar] [CrossRef]
  30. Davis, J. Consideration of atmospheric turbulence in laser systems design. Appl. Opt. 1966, 5, 139–147. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the proposed method.
Figure 1. Schematic diagram of the proposed method.
Sensors 23 08737 g001
Figure 2. Laser spot diagram. (a) Single original laser spot; (b) single received laser spot; (c) multiplex original laser spot; (d) multiplex received laser spot.
Figure 2. Laser spot diagram. (a) Single original laser spot; (b) single received laser spot; (c) multiplex original laser spot; (d) multiplex received laser spot.
Sensors 23 08737 g002
Figure 3. Intensity images of received LG beam. (a) | l 0 | = 1 , 2 , 10 (single OAM). (b) | l 0 | = 1 , 2 , 8 (multiplexed OAM).
Figure 3. Intensity images of received LG beam. (a) | l 0 | = 1 , 2 , 10 (single OAM). (b) | l 0 | = 1 , 2 , 8 (multiplexed OAM).
Sensors 23 08737 g003
Figure 4. Spatial distribution for a single topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 . (b) C n 2 = 1 × 10 15   m 2 / 3 .
Figure 4. Spatial distribution for a single topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 . (b) C n 2 = 1 × 10 15   m 2 / 3 .
Sensors 23 08737 g004
Figure 5. RMSE of the five methods for single topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Figure 5. RMSE of the five methods for single topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Sensors 23 08737 g005
Figure 6. G-mean of the five methods for single topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Figure 6. G-mean of the five methods for single topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Sensors 23 08737 g006
Figure 7. Spatial distribution for multiplexed topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Figure 7. Spatial distribution for multiplexed topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Sensors 23 08737 g007
Figure 8. RMSE of the five methods for multiplexed topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Figure 8. RMSE of the five methods for multiplexed topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Sensors 23 08737 g008
Figure 9. G-mean of the five methods for multiplexed topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Figure 9. G-mean of the five methods for multiplexed topological charges dataset. (a) C n 2 = 1 × 10 14   m 2 / 3 ; (b) C n 2 = 1 × 10 15   m 2 / 3 .
Sensors 23 08737 g009
Table 1. Confusion matrix.
Table 1. Confusion matrix.
Actual ResultPredicted Result
PositiveNegative
PositiveTP (True Positive)FN (False Negative)
NegativeFP (False Positive)TN (True Negative)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, H.; Chen, C.; Hu, X.; Yang, H. An Efficient Recognition Method for Orbital Angular Momentum via Adaptive Deep ELM. Sensors 2023, 23, 8737. https://doi.org/10.3390/s23218737

AMA Style

Yu H, Chen C, Hu X, Yang H. An Efficient Recognition Method for Orbital Angular Momentum via Adaptive Deep ELM. Sensors. 2023; 23(21):8737. https://doi.org/10.3390/s23218737

Chicago/Turabian Style

Yu, Haiyang, Chunyi Chen, Xiaojuan Hu, and Huamin Yang. 2023. "An Efficient Recognition Method for Orbital Angular Momentum via Adaptive Deep ELM" Sensors 23, no. 21: 8737. https://doi.org/10.3390/s23218737

APA Style

Yu, H., Chen, C., Hu, X., & Yang, H. (2023). An Efficient Recognition Method for Orbital Angular Momentum via Adaptive Deep ELM. Sensors, 23(21), 8737. https://doi.org/10.3390/s23218737

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop