Abstract
The devastating consequences of climate change have resulted in the promotion of clean energies, being the wind energy the one with greater potential. This technology has been developed in recent years following different strategic plans, playing special attention to wind generation. In this sense, the use of bicomponent materials in wind generator blades and housings is a widely spread procedure. However, the great complexity of the process followed to obtain this kind of materials hinders the problem of detecting anomalous situations in the plant, due to sensors or actuators malfunctions. This has a direct impact on the features of the final product, with the corresponding influence in the durability and wind generator performance. In this context, the present work proposes the use of a distributed anomaly detection system to identify the source of the wrong operation. With this aim, five different one-class techniques are considered to detect deviations in three plant components located in a bicomponent mixing machine installation: the flow meter, the pressure sensor and the pump speed.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
European Union (EU) is committed to generate a new ecological transition to become in the world’s first climate-neutral continent by 2050. This effort requires a significant investment from public and private entities that EU tries to impulse through its European Green Deal’s Investment Plan [1]. In this sense, renewable energies play a key role in this plan to fight against the fossil fuel markets. Among the different options, wind power is the most extended in EU [2]. According to [3], in 2018, the global electricity from renewable sources represented only a 25.6 % of the total share. This value reveals a significant development in comparison with previous years [4, 5]. Although, nowadays, the hydroelectric energy is the most common, covering 16 % of the generation, the wind technologies have experienced a remarkable increase during the last two decades. In [6], it is stated that the wind power installed grew from 17 GW in 2000 to 591 GW in 2018. According to this sequence, there are some forecasts that estimate a global share of 23 % in 2030 [7]. This represents four to five times the current share [8]. In this context, there are several researches oriented to optimise and improve different renewable energies systems [9]. Specially, in case of wind turbines, it takes significant relevance the wind blades in terms of shape [10] and materials [11, 12].
Most of the commercial wind blades are made of polyester or epoxy reinforced with glass fibre or carbon fibre. During the manufacturing process, a key step occurs when the epoxy is mixed with catalytic to obtain the resin for the infusion process [13]. Due to the non-Newtonian nature of both fluids, this is not a trivial step, since the mix of both fluids must kept in the same proportion for any flow demanded, and the presence of noise during monitoring can hinder the control of the system.
The facility under study obtains carbon fibre as final product, whose features in terms of stiffness, tensile and compression strengths are suitable to the use in wind generator blades. It is an interesting alternative to glass fibre, since it shows greater stiffness with lower density than the glass fibres, resulting in lighter blades. However, an accurate proportion of both bicomponents should be ensured to avoid possible wrinkles that could lead to undesired material features [14]. Then, flow rate deviations could have an important impact on material, such as product waste, defective material during curing process, or shortening of blade life. Both scenarios result in economic losses.
Therefore, an early diagnosis of any deviation from the normal operation of the system is crucial to take actions in several directions. On the one hand, the anomaly detection can be used to carry out corrective maintenance, and on the other hand, it can be used to develop fault tolerant systems. This means that the control system can continue the process trying to avoid the propagation of the anomaly effect. In both cases, besides the detection of an unexpected event, it is interesting to determine the part of the system that has suffered the fault. Then, it is desirable to develop a distributed topology, where an anomaly detection system is placed over each sensor or actuator. The left side of Fig. 1 represents the implementation of an anomaly detection system applied to the entire mixing facility, and right side depicts a distributed topology. This second topology has the advantage of locating the source of the anomaly.
In this sense, anomaly detection techniques [15] provide an important advantage during monitoring process to identify the abnormal behaviour of the system. From a generic approach, an anomaly can be considered as an unusual pattern that does not an expect behaviour [16]. The used techniques for anomaly detection can be categorised in three main types, based on type of learning methods used:
-
Type 1: Supervised Learning: in this case, all samples are labelled as normal and anomalous. Then, modelling techniques are applied over each class.
-
Type2: Unsupervised Learning: here, there is not a previous knowledge about the dataset and therefore samples are not labelled.
-
Type 3: Semi-supervised Learning: in this case, only samples belonging to normal operation are used to train the models, so just the normal operation is modelled.
Given the fact that in most cases, only information from correct operation is available, the definition of this circumstance is mandatory to determine an anomalous situation [16, 17]. Then, it is defined the concept of one-class classifiers that cover all the instances that belong to the target set, which is the set representing the normal operation. The use of one-class classifiers has been widely used for many different applications, such as anomaly detection, fraud detection or medicine diagnosis. However, this semi-supervised approach offers only the possibility of labelling data as target or non-target data. In [18], a one-class classifier is implemented to determine the occurrence of anomalies in a bearing system. Although it is an interesting tool, it is not capable of locating or, at least, narrowing the source of the anomaly. Focusing on a different approach, in [19], a virtual sensor using intelligent techniques is developed to determine the value measured by a sensor located in an industrial plant. Then, the deviation between real and predicted value is the criteria to determine anomalies. This approach can be considered to identify the source of the anomaly, and thus favour the diagnostic and supervision of the plant.
This work faces the anomaly detection in an industrial plant used to obtain bicomponent material used for wind generator blades. The proposal takes advantage of one-class techniques to implement a distributed system capable of locating anomalies in specific parts of the facility. To achieve this goal, a real dataset gathered during the correct operation with different operating points is considered. Then, a deep analysis of different one-class techniques is carried out to evaluate the anomaly detection using three different plant anomalies. The anomalies are created by deviating the values measured with a flow meter, the pressure and the speed in a pump.
The present research is organised as follows. After this contextualisation section, the case of study section presents the bicomponent mixing machine. Later, Sect. 3 introduces the theoretical concepts of the four one-class anomaly detection techniques used in this work, to present in Sect. 4 the experiments and results and finally derive the conclusions and future works in Sect. 5.
2 Case study
In this section, the bicomponent mixing machines are described, detailing its main components and explaining its workflow. Also, the recorded dataset is described in terms of measured variables and sampling procedure, presenting an example of the dataset graphically.
2.1 Bicomponent mixing machine for turbine blade material
During fabrication of a turbine blade, the shape of the blade is created by combining a resin with reinforcement of fibre glass or carbon. In order to produce such resin, it is necessary an industrial machine to mix two primary fluids, the epoxy and the catalytic to obtain a final product which presents high tensile and compressive strengths and great chemical resistance [20]. This industrial mixing machine is schematised in Fig. 2.
As can be seen in Fig. 2, the system is monitored by means of a total of 9 sensors. Each fluid line has one pressure sensor at the output of the pumps (\(P_{E1}\), \(P_{C1}\)) and another at the input of the mixing valve (\(P_{E2}\), \(P_{C2}\)) and also one flow meter per line (\(F_{E2}\), \(F_{C2}\)). Each pump has a speed sensor (\(S_{E}\), \(S_{C}\)), and finally, the flow of mixed material at the output line is also measured (\(F_{M}\)).
The workflow is as follows:
-
1.
Both fluids are stored in separated tanks and sent to the mixing valve by centrifugal pumps controlled through three phase variable frequency drives (VFD).
-
2.
A control system adjusts the pumps speed, acting over each VFD, based on information sampled in the three flow meters to keep a constant output flow and homogeneous material.
2.2 Dataset
With the aim of generating an accurate anomaly detection model, the bicomponent mixing process is monitored by using a total of nine sensors, presented in previous Section. The monitoring process has a sample period 0.5 sec during an hour and 15 min. After removing wrong measures (null or negative values), the final dataset consists of 8549 samples and nine variables.
In order to illustrate how sample values varies in time, four of the measures are presented in Fig. 3 during 8 min.
As all data samples represents the normal operation of the system, anomalies have been generated following the process shown in Fig. 4.
Here, j anomalies were generated by modifying in a \(p\%\) one of the random variables in a total of j random samples from a MxN dimensional dataset.
3 Soft computing methods to validate the proposal
This section details the one-class techniques used to detect anomalies over the case of study previously presented. All algorithms work only with samples from normal behaviour of the system, and later anomalies are used to evaluate the performance of each one.
3.1 k-nearest neighbour
The k-Nearest Neighbour (kNN) for anomaly detection is based on relative distances between neighbour training samples and new test samples, therefore the local density of the hypersphere made of its \(k^{th}\) nearest neighbours [21], determines whether a sample x is considered as an anomaly or not.
It can be expressed by Eq. (1) that calculates such density d, as the relation between the distance of a new testing sample \(\varvec{x}\) to the \(k^{th}\) nearest training data neighbour \(kNN^{tr}(\varvec{x})\) and the local distance from the \(k^{th}\) neighbour to its \(k^{th}\) neighbour [17]
Figure 5 represents both distances graphically in a easier way.
The value of k plays a significant role in the classifier performance, and it depends on the dataset structure shape [21].
3.2 Minimum spanning tree
This one-class method is based on target set modelling by means of the structure obtained by a minimum spanning tree (MST). It relies on the assumption that two points \(p_i, p_j \in \mathbb {R}^{n}\) that belong to target class should be neighbours in \(\mathbb {R}^{n}\) representation [22]. Then, a linear transformation can be found for these points and all points considered as target class. As this set commonly contains more than two objects, then, more than one transformation can be applied. Hence, for a dataset D with n instances, \((n-1)\) linear transformations can be found [22]. Then, the MST consists of a set of edges \(e_{ij}\) that specify the linear transformations of each point. It is implemented a graph ensuring the absence of loops and minimising the total length of the edges. It can be defined as a problem of finding \((n-1)\) edges that implement a tree with a minimum total weight, as shown in Eq. (2) and subjected to Eq. (3).
Once the MST is trained with data from the target set, the distance between a new test object \(\varvec{p}\) to training set D, which is also the target set, is calculated as the shortest distance to the set of \((n-1)\) edges of the tree, as shown in Eq. (4). If this distance exceeds a threshold, \(\beta\) can be derived from a quantile function of the obtained MST.
3.3 Non-convex boundary over projections
Non-Convex Boundary over Projections (NCBoP) is a novel one-class classification algorithm that overcome the weaknesses presented by well-known Convex Hull over random projections [23, 24].
The main basis of this method is to approximate the boundaries of a dataset \(S \in \mathbb {R}^{n}\) using the non-convex hull over \(\pi\) random projections on 2D planes and then, determine the non-convex limits on that plane, reducing in this way the complexity of calculating the non-convex limits over \(\mathbb {R}^{n}\).
The process to determine the non-convex polygon that contains a set of points \(D = {d_0,d_1,...,d_i}\) follows the next steps:
-
Look for the point with the lowest y-coordinate (\(d_0\)). In case tie, the point with the lowest x-coordinate is selected.
-
Find the k-nearest points to that current point. For this, the vectors \(d_0 d_i\) with \(i\in \{1, \cdots , n\}\) are implemented, looking for the smallest Euclidean distance.
-
Sort the k-nearest points based on the angle resulting by the line with the x-axis. The further point from the starting point is selected.
-
Once this first couple of points are calculated, a stack structure is created including the next third point.
-
Then, it is checked if the next point in the list turns left (pushed into the stack) or right (point on the top is removed). This process is repeated until arrive again to the starting point.
Once the non-convex polygon is defined, it should have all training instances. Finally, a parameter \(\lambda\) is considered to increase or decrease the polygon vertexes v of the non-convex polygon NC(X) from the centroid c, according to Eq. (5).
Then, when a testing point is presented to the algorithm, if such point falls outside the non-convex polygon in at least one of the projections, it is considered as an anomaly, otherwise it represents the normal behaviour of the system. Figure 6 presents a 3D example where a sample that falls outside the limits of the non-convex polygon for one the calculated projections.
3.4 Principal component analysis
The principal component analysis [25] can be described as a linear operation involving unsupervised learning for data compression whose aim is to find that orthogonal basis that maximises the data variance for a given dimension of basis.
The final goal is to generate a new subspace using as new components the eigenvalues of the covariance matrix of the original input space. From a geometrical point of view, it consists on a rotation of the original axes, which can be expressed by Eq. (6).
where \(x^{d}\) in an N-dimensional space onto vectors in an M-dimensional space \((x_{1}...x_{N})\), where \(M\le N\), \(W_i\) are the N eigenvectors of the covariance matrix, and \(y_i\) are the projected original data onto the new output M-dimensional subspace \((y_{1}...y_{M})\).
Finally, to the criteria applied to identify an anomaly [26,27,28], consists of calculating the reconstruction error and determine if this reconstruction error is greater than the maximum threshold. The reconstruction error is expressed in Eq. (7) as the difference between the original dataset and its projection over the new subspace Eq. (8), computed in the original data. Such projection can be expressed as:
This technique offers good results when the subspace is clearly linear [17].
3.5 Support vector data description
The support vector machine (SVM) is a supervised learning used for classification and regression tasks [21]. Its main goal is to map the training set into a hyperspace and, then, implement a hyperplane to maximises the distance between classes [29].
Support vector data description (SVDD) aims to find a minimum volume space that contains all training data samples. Therefore, given a dataset X in \(\mathbb {R}^{n}\), SVDD will try to find the hypersphere with centre a and radius R that includes most of the training data samples. It can be formulated as an optimisation problem with Eqs. (9) and (10).
Due to the possible presence of noise in the dataset, the parameter \(\xi\) is used to control the slack in the limits of the hypersphere. Parameter C is also considered to adjust the volume-errors trade-off.
When a new testing sample is presented to the trained SVDD, the distance from this new sample to the centre of the hypersphere is calculated and if this distance is greater than radius R, it is considered as an outlier.
4 Experiments and results
4.1 Experiments setup
The performance of each one-class technique was evaluated according to the different configurations described next:
-
kNN:
-
Number of neighbours (NN) = 1:5:25.
-
Fraction of anomalies in the target set (\(\nu\)): 0 : 0, 05 : 0, 25.
-
-
MST:
-
Fraction of anomalies in the target set (\(\nu\)): 0 : 0, 05 : 0, 25.
-
-
NCBoP:
-
Number of projections \(\pi\) = 5, 10, 50, 100, 500, 1000.
-
Expansion parameter \(\lambda\) = 0, 5 : 0.1 : 1, 5.
-
-
PCA:
-
Fraction of anomalies in the target set (\(\nu\)): 0 : 0, 05 : 0, 25.
-
Components \(\beta\) = 1 : 1 : 9.
-
-
SVDD:
-
Fraction of anomalies in the target set (\(\nu\)): 0 : 0, 05 : 0, 25.
-
Kernel width (%) = 0.1 : 0, 1 : 10.
-
The process followed to evaluate the performance of each classifier is based on the \(k-fold\) cross-validation method, with \(k = 5\), which is depicted in Figure 7. This procedure consists of dividing in a random way the target set in five different groups. Then, five classifiers are implemented using the 80 % of the data and leaving the 20 % to the test phase. The use 5 folds ensures that all instances are considered for the training and test phases. Then, the classifier is tested using the non-target set and 20 % of the target data.
Furthermore, for each of the configurations tested, the data followed three different conditioning prior the classifier training:
-
Type A: normalisation between the interval 0 to 1.
-
Type B: normalisation using the mean and standard deviation of each variable. This is also known as ZScore normalisation [30].
-
Type C: the data remains unaltered without any pre-processing technique.
The performance of each classifier configuration is evaluated using the well-known area under the receiving operating characteristics curve (AUC) measure [31]. The AUC combines the true positive and false positive rates, obtaining a unique measure of the classifier performance. From a statistical point of view, this value represents the probability of classifying a random positive instance as positive [31]. Furthermore, in contrast to other parameters such as sensitivity, precision or recall, AUC is not sensitive to class distribution, which is a significant advantage especially in one-class tasks [32]. This parameter is calculated five times, one for each fold, and its mean value and standard deviation between runs (STD, in %) are also registered. As a comparative analysis of each technique is sought, the time needed to train (\(t_{tr}\)) each classifier is also registered as a measure of the computational cost of the classifiers.
To evaluate the proposal of a distributed one-class classifier for detecting anomalies in different plant components, three different variables are deviated from its correct operation:
-
Flow rate of epoxy resin (FE).
-
Speed of catalyst pump (SE).
-
Pressure measure in epoxy resin pump (PE1).
These anomalies are generated by modifying the original measurements a \(\pm 5 \%\), \(\pm 10 \%\), \(\pm 15 \%\), \(\pm 20 \%\) and \(\pm 25 \%\).
4.2 Results
This subsection describes the best results, in terms of AUC for each technique, component with anomalous behaviour, and percentage deviation. This information is reflected in Tables 1, 2 and 3. Furthermore, to provide a deeper insight of the results, the behaviour of each technique depending on the deviated percentage is presented in Figs. 8, 9 and 10.
Focusing on the anomaly appearance of the flow meter sensor, it is important to remark that PCA is in almost all cases the technique with the best AUC value, with at least 96,22 %. These results are achieved with a significantly high number of components, varying between 8 and 9. In these experiments, MST and kNN compete with PCA in terms of AUC. However, they present a remarkable disadvantage comparing with PCA, which is the training time, whose values are more than one hundred times greater. NCBoP does not reach the level of PCA, MST and kNN, but in all cases exceeds a minimum of 87,62 % AUC. Finally, SVDD shows a gradually increase in the anomaly detection task as the percentage deviation grows, starting with a 61,97 % for a 5% deviation and finishing with a 98,23 % when the deviation is 25 %. Furthermore, SVDD is the unique technique that offer the best results with ZScore normalisation.
Regarding the detection of anomalous speed measurements, the obtained results are slightly different than the ones obtained with the flow meter deviations. In this case, NCBoP and MST present the best performance in terms of AUC. However, NCBoP has a computation time that is significantly greater than the presented by MST. PCA and kNN show similar performance, especially with 15 %, 20 % and 25 % deviation. It is interesting to emphasise the fact that, in this case, the number of components \(\beta\) is significantly lower than the obtained in first experiment. The trend shown by SVDD in previous experiment is repeated in this case, since the AUC increases gradually with the percentage variation.
In respect of the third experiment, in which the pressure measured is modified, PCA, MST and NCBoP compete again for the best results, being PCA the fastest technique. This is a key issue to select this technique in case of a tie. SVDD is once again the worst technique taking AUC and training time values. It is interesting to indicate that it is, as it happened in the two previous experiments, the unique technique whose best performance is achieved with ZScore pre-processing. Finally, it worth mentioning that the number of neighbours is one in all cases but one.
5 Conclusions and future works
The present work presents five different one-class classifier techniques to detect anomalies in three components of an industrial system used to obtain a bicomponent material. These anomalies are artificially generated by deviating the original measurement a variable percentage, modifying the flow, pressure and speed measured. From the experiments developed and the obtained results, in general terms, it can be assumed that, in the three experiments, the one-class techniques presented a successful performance. A valuable contribution of this work consists of the implementation of three one-class classifiers to consider the possibility locating the source of the anomaly. Hence, when the classifier assigned to the flow meter detects a deviation, it can be used to isolate its measurements and avoid its propagation. Besides this application, the idea of incorporating this distributed topology could reduce significantly the maintenance cost, since early detection of wrong performance can be addressed.
In future works, it could be considered the use of imputation techniques or intelligent models to recover the real data when it is lost due to wrong sensor performance. Then, a combination of system modelling and distributed one-class topology would impact in the system optimisation. Furthermore, an online training stage as the system is in operation would improve the classifier performance. Finally, the possibility of applying a prior clustering process to obtain hybrid intelligent classifiers can be considered. Hence, each cluster would correspond to a different operating point.
References
European Commission (2021). A European Green deal. https://ec.europa.eu/info/strategy/priorities-2019-2024/european-green-deal_en. Accessed 18 February 2021
Eurostat (2021). Renewable energy statistics. https://ec.europa.eu/eurostat/statistics-explained/index.php/Renewable_energy_statistics#Wind_and_water_provide_most_renewable_electricity.3B_solar_is_the_fastest-growing_energy_source. Accessed 18 February 2021
Repsol’s Economic Research Department (2020) Annual energy-statistics 2020. https://www.repsol.com/content/dam/repsol-corporate/en_gb/energia-e-innovacion/annual-energy-statistics-2020_tcm14-168076.pdf. Accessed 4 Nov 2021
Owusu PA, Asumadu-Sarkodie S (2016) A review of renewable energy sources, sustainability issues and climate change mitigation. Cogent Eng 30(1):1167990
Lund H (2007) Renewable energy strategies for sustainable development. Energy 32(6):912–919
Asociación Empresarial Eólica (2019) Anuario eólico. La voz del sector 2019. https://www.aeeolica.org/images/Publicaciones/Anuario-Elico-2019.pdf. Accessed 15 Oct 2021
Infield D, Freris L (2020) Renewable energy in power systems. Wiley, Chichester
BP plc (2019). Renewable energy - wind energy. https://www.bp.com/en/global/corporate/energy-economics/statistical-review-of-world-energy/renewable-energy.html.html#wind-energy. Accessed 3 Mar 2021
Zuo Y, Liu H (June 2012) Evaluation on comprehensive benefit of wind power generation and utilization of wind energy. In: Software engineering and service science (ICSESS), 2012 IEEE 3rd international conference on. pp 635–638
Xudong W et al (2009) Shape optimization of wind turbine blades. Wind Energy Int J Progr Appl Wind Power Conv Technol 12(8):781–803
Cognet V, du Pont SC, Thiria B (2020) Material optimization of flexible blades for wind turbines. Renew Energy 160:1373–1384
Brondsted P, Nijssen RP (Eds) (2013) Advances in wind turbine blade design and materials. Woodhead Publishing Limited
Bank LC, Arias FR, Yazdanbakhsh A, Gentry TR, Al-Haddad T, Chen JF, Morrow R (2018) Concepts for reusing composite materials from decommissioned wind turbine blades in affordable housing. Recycling 3(1):3
Mishnaevsky L, Branner K, Petersen HN, Beauson J, McGugan M, Sorensen BF (2017) Materials for wind turbine blades: an overview. Materials 10(11):1285
Miljković D (2011) Fault detection methods: a literature survey. In: MIPRO, 2011 proceedings of the 34th international convention. pp 750–755. IEEE
Chandola V, Banerjee A, Kumar V (2009) Anomaly detection: a survey. ACM Comput Surv (CSUR) 41(3):15
Tax DMJ (2001) One-class classification: concept-learning in the absence of counter-examples [ph. d. thesis]. Delft University of Technology
Zeng M, Yang Y, Luo S, Cheng J (2016) One-class classification based on the convex hull for bearing fault detection. Mech Syst Signal Process 81:274–293
Jove E, Casteleiro-Roca JL, Quintián H, Méndez-Pérez JA, Calvo-Rolle JL (2019) Virtual sensor for fault detection, isolation and data recovery for bicomponent mixing machine monitoring. Informatica 30(4):671–687
Jove E, Casteleiro-Roca J, Quintián H, Méndez-Pérez JA, Calvo-Rolle JL (2020) Anomaly detection based on intelligent techniques over a bicomponent production plant used on wind generator blades manufacturing. Revista Iberoamericana de Automática e Informática industrial 17(1):84–93
Sukchotrat T (2008) Data mining-driven approaches for process monitoring and diagnosis. Doctoral dissertation, The University of Texas at Arlington
Juszczak P, Tax DM, Pe E, Duin R (2009) Minimum spanning tree based one-class classifier. Neurocomputing 72(7–9):1859–1869
Casale P, Pujol O, Radeva P (2011) Approximate convex hulls family for one-class classification. In: International workshop on multiple classifier systems. pp 106–115. Springer
Fernandez-Francos D, Fontenla-Romero O, Alonso-Betanzos A (2017) One-class convex hull-based algorithm for classification in distributed environments. IEEE Trans Syst Man Cybern Syst 50(2):386–396
Pearson K (1901) On lines and planes of closest fit to systems of points in space. Philos Mag 2(6):559–572
Mazhelis O (2006) One-class classifiers: a review and analysis of suitability in the context of mobile-masquerader detection. South African Comput J 2006:29–48
Chiang LH, Russell EL, Braatz RD (2000) Fault detection and diagnosis in industrial systems. Springer Science & Business Media, Heidelberg
Wu J, Zhang X A pca classifier and its application in vehicle detection. In: IJCNN’01. International joint conference on neural networks. Proceedings (Cat. No. 01CH37222), volume 1, IEEE, pp 600–604
Rebentrost P, Mohseni M, Lloyd S (2014) Quantum support vector machine for big data classification. Phys Rev Lett 113:130503. https://doi.org/10.1103/PhysRevLett.113.130503
Shalabi LA, Shaaban Z (May 2006) Normalization as a preprocessing engine for data mining and the approach of preference matrix. In: 2006 International conference on dependability of computer systems. pp 207–214
Fawcett T (2006) An introduction to ROC analysis. Pattern Recogn Lett 27(8):861–874
Bradley AP (1997) The use of the area under the roc curve in the evaluation of machine learning algorithms. Pattern Recogn 30(7):1145–1159
Acknowledgements
CITIC, as a Research Center of the University System of Galicia, is funded by Consellería de Educación, Universidade e Formación Profesional of the Xunta de Galicia through the European Regional Development Fund (ERDF) and the Secretaría Xeral de Universidades (Ref. ED431G 2019/01).
Funding
Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
Authors do not declare any conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Zayas-Gato, F., Michelena, Á., Jove, E. et al. A distributed topology for identifying anomalies in an industrial environment. Neural Comput & Applic 34, 20463–20476 (2022). https://doi.org/10.1007/s00521-022-07106-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-022-07106-7