Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
GEANT4 Simulation of the Gamma-Ray Total Absorption Facility
Previous Article in Journal
Effective Text Classification Through Supervised Rough Set-Based Term Weighting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Delamination Prediction in Layered Composites Using Optimized ANN Algorithms: A Comparative Analysis

Faculty of Aeronautics and Astronautics, Istanbul Technical University, Istanbul 34467, Turkey
Symmetry 2025, 17(1), 91; https://doi.org/10.3390/sym17010091
Submission received: 9 December 2024 / Revised: 4 January 2025 / Accepted: 7 January 2025 / Published: 9 January 2025
(This article belongs to the Section Engineering and Materials)

Abstract

:
This study investigates the effectiveness of Artificial Neural Networks (ANNs) in predicting the outcomes of Double Cantilever Beam (DCB) tests, focusing on time and force as input variables and displacement as the predicted output. Three ANN training algorithms—Scaled Conjugate Gradient (SCG), Broyden Fletcher Goldfarb Shanno (BFGS) Quasi-Newton, and Levenberg-Marquardt (LM)—were evaluated based on prediction accuracy and computational efficiency. A parametric study was performed by varying the number of neurons (from 10 to 100) in a single hidden layer to optimize network structure. Among the evaluated algorithms, LM demonstrated superior performance, achieving prediction accuracies of 99.6% for force and 99.3% for displacement. In contrast, SCG exhibited the fastest convergence but had a significantly higher error rate of 8.6%. The BFGS algorithm provided a compromise between accuracy and speed but was ultimately outperformed by LM in terms of overall precision. In addition, configurations with up to 100 neurons were tested, indicating that although slightly lower error rates could be achieved, the increase in computation time was substantial. Consequently, the LM algorithm with 50 neurons delivered the best balance between accuracy and computational cost. These findings underscore the potential of ANNs, particularly LM-based models, to enhance material design processes by providing reliable predictions from limited experimental data, thereby reducing both resource utilization and the time required for testing.

1. Introduction

Artificial intelligence (AI) applications have become increasingly prevalent across diverse sectors, providing substantial gains in efficiency and accuracy. Although traditional iterative techniques can be effective, they often demand considerable computational resources and time, and their reliability may be compromised without a robust control system—posing significant challenges for data validation. In contrast, AI offers a more agile and efficient alternative, making it especially valuable for complex problems where conventional methods frequently encounter limitations [1].
In materials science, development processes are typically trial-based and require extensive experimentation, which is both costly and resource-intensive. Implementing AI systems offers an opportunity to streamline these processes by reducing the reliance on physical testing. AI’s minimal computational cost, self-regulating capabilities, and rapid decision-making render it well-suited for trial-based approaches in material development [2,3]. Vasudevan et al. [4] underscored the progress AI has stimulated in this field, illustrating how integrating computational models, large-scale experiments, and imaging techniques can expedite the innovation of novel material systems. Accordingly, materials science appears to be moving toward a data-driven future, in which AI significantly enhances the pace of understanding and discovering new materials.
Among various AI methodologies, Artificial Neural Networks (ANNs) have emerged as a particularly powerful tool due to their improved computational efficiency and capacity to provide fast, reliable predictions. ANNs have been widely utilized in multiple aspects of material development [5,6,7,8,9]. For instance, Kang et al. [5] applied backpropagation neural networks to predict toughness values in structural steels based on tensile test data, examining the influence of crack plane orientation and temperature. Similarly, Guan et al. [6] harnessed an ANN to estimate fracture toughness in Nb–silicide composites by inputting both microstructural characteristics and material composition. In another example, Zakaulla et al. [7] developed a multi-layer perceptron network to model the mechanical properties of PEEK/Carbon/Titanium composites. Genel [8] and Kazi et al. [9] also demonstrated that ANNs can accurately predict load and displacement from experimental data.
In this study, Artificial Neural Networks (ANNs) were chosen for optimization due to their demonstrated effectiveness in capturing complex, non-linear relationships—particularly for multi-dimensional force–displacement data in Mode I delamination. While certain traditional methods (e.g., gradient-based or heuristic techniques) can involve specific assumptions or parameter constraints, ANNs offer flexibility by adapting to diverse experimental data without requiring extensive prior knowledge. They also scale well, allowing for architectures (e.g., hidden layer size) to be tailored to the problem at hand and enabling computational efficiency once trained, which can facilitate rapid iterative design. Moreover, existing research suggests that ANNs—especially those employing the Levenberg–Marquardt (LM) algorithm—can, under suitable conditions, achieve high accuracy and convergence speed for experimental data modeling [10,11,12]. Their robust tolerance to noise enhances reliability even in the face of experimental inconsistencies, making them particularly well-suited for modeling composite material behavior. Collectively, these advantages help minimize computational time and maximize prediction accuracy, thereby highlighting ANNs as a strong candidate for optimizing delamination predictions in composite materials.
The aerospace and automotive industries heavily rely on lightweight, high-strength composite materials; however, their susceptibility to delamination—especially under impact loads—remains a primary challenge. This failure mode can significantly affect overall strength and fatigue life. Theoretical frameworks, such as Classical Laminate Theory (CLT) and First-Order Shear Deformation Theory (FSDT), guide the design of laminate structures to mitigate delamination risks. Delamination failures are commonly classified into Mode I (caused by normal or opening loads) and Mode II (driven by interlaminar shear stresses) [13]. ANNs have shown promise in predicting the size, location, and shape of delamination, thus contributing to a deeper understanding of composite failure [14,15,16,17,18,19]. For example, Bortoluzzi et al. [17] illustrated the efficacy of using an ANN trained via the Levenberg–Marquardt algorithm to predict Mode II interlaminar fracture toughness in z-pinned composites. Likewise, Sharma et al. [19] trained and validated an ANN model using impact test data to estimate the crack initiation toughness in polymer composites. Allegri [18] compared a semi-empirical power-law model with a single hidden-layer ANN model for predicting fatigue delamination growth, noting that the ANN achieved high accuracy despite its relatively simple structure.
Building on these findings, the present study employs force–time data from Mode I delamination tests as inputs to train and validate an ANN model, with displacement values serving as the predicted output. Three algorithms—Scaled Conjugate Gradient (SCG), Broyden–Fletcher–Goldfarb–Shanno (BFGS) Quasi-Newton, and Levenberg–Marquardt (LM)—are compared in terms of both predictive accuracy and computational efficiency. Additionally, the number of neurons in a single hidden layer veries from 10 to 100 to optimize network performance. Ultimately, this research aims to generate artificial experimental data by modeling displacement values based on force–time inputs, facilitating the estimation of composite fracture toughness and enabling validation against real-world experimental outcomes.
Recent studies have demonstrated the application of ANNs in composite materials research, addressing various challenges associated with delamination prediction. For instance, Sreekanth et al. [16] utilized vibration signals to locate and quantify delamination in glass fiber-reinforced polymer (GFRP) plates, employing a supervised feed-forward ANN. This approach showed accurate predictions but was limited by its dependency on pre-selected signal features, which might not generalize well across varying composite systems. Similarly, Oliver et al. [20] integrated frequency shifts as inputs to an ANN for detecting and quantifying damage in composite-laminated plates, achieving high diagnostic accuracy yet focusing primarily on damage identification rather than mechanical properties. In a more advanced methodology, Rautela et al. [21] explored unsupervised feature learning through deep convolutional autoencoders to identify anomalies in composite panels. Their work highlighted the potential of deep learning approaches for complex data patterns but faced computational challenges associated with large-scale training datasets.
The present study uniquely contributes to this field by systematically comparing three ANN algorithms—Scaled Conjugate Gradient (SCG), Broyden–Fletcher–Goldfarb–Shanno (BFGS), and Levenberg–Marquardt (LM)—specifically for Mode I delamination toughness prediction. Unlike previous works, this approach focuses on balancing prediction accuracy and computational efficiency. The LM algorithm, optimized with 50 neurons in a single hidden layer, emerged as the most effective, achieving prediction accuracies of 99.6% for force and 99.3% for displacement. These results surpass the accuracy metrics reported in recent works and provide a robust alternative for resource-efficient data-driven material design. Additionally, while Mukherjee and Routroy [22] demonstrated the superiority of the LM algorithm in minimizing errors for industrial processes, the present study advances this understanding by applying the LM algorithm specifically to composite material delamination prediction. This highlights its adaptability and effectiveness in a domain with intricate mechanical behaviors. Moreover, the systematic analysis of computational cost versus accuracy trade-offs presented in this study offers a novel perspective for ANN model optimization, which is less explored in the existing literature.

2. Experimental Study

2.1. Fabrication of Composite Specimens

Unidirectional carbon fibers (CF) (12k fabric, 300 g/m2) were used as the reinforcement, and Hexion MGSL160 epoxy was combined with its hardener at a 100:25 weight ratio to form the thermoset polymer matrix. Prior to vacuum infusion, the epoxy–hardener mixture was stirred for three minutes in the specified ratio, and any trapped air bubbles were removed using a desiccator. A vacuum pump then injected the resin into the mold at a unidirectional flow rate of 50 mm/min to ensure adequate fiber wetting.
Twelve unidirectional carbon fiber mats were stacked to create notched double cantilever beam (DCB) test specimens. A 10 mm-thick Teflon film was inserted at the mid-plane of the laminate to introduce an initial crack of 50 mm length for the DCB tests. The fiber volume fraction was determined in accordance with ASTM D3171 [23] using five samples (10 × 20 mm2) from each laminate, yielding an average fiber volume fraction of 55.57%.
Following fabrication, the laminates were cut to dimensions of 150 × 21 × 3.80 mm by water jet, and the loading blocks were affixed to the specimen ends, as illustrated in Figure 1.

2.2. Test Methodology

Double Cantilever Beam (DCB) tests were conducted at room temperature using an MTS™ Series 322 testing system. A displacement rate of 2 mm/min was applied in accordance with ASTM D5528/D5528M [25] to obtain the force–displacement curves. During testing, a digital camera was used to record delamination progression, facilitating the calculation of Mode I delamination toughness.
Three specimens were tested under these conditions, and force and displacement data were recorded until the delamination reached the specimen’s end. Using the force–displacement measurements and the observed delamination length, the Mode I delamination toughness (GIc) was determined as shown in Equation (1) [25]:
G I c = m P c δ c 2 b a F N ,
where Pc is the applied load, δc is the corresponding displacement, b is the specimen width, a is the delamination length, F is a correction factor for significant deflections, and N is a factor that accounts for large displacements and fracture. The parameter m is obtained via a linear least-squares regression of the normalized compliance log(C/N) versus log(a), where the compliance C is defined as δ/P.
The correction factors F and N were calculated according to Equations (2) and (3), respectively [25]:
F = 1 3 10 δ c a 2 3 2 δ c t a 2 ,
N = 1 L a 3 9 8 1 L a 2 δ c t a 2 9 35 δ c a 2 ,
where t and L′ are defined as shown in Figure 1.
The experimental setup was carefully designed to ensure balance and consistency in the data, reflecting symmetry-like properties. This approach minimizes potential biases and ensures that the force displacement relationships are accurately captured for ANN modeling.

3. ANN Model and Training

In this study, Artificial Neural Networks (ANNs) were employed to predict Mode I delamination behavior in composite materials. The training process utilized three optimization algorithms—Scaled Conjugate Gradient (SCG), Broyden–Fletcher–Goldfarb–Shanno (BFGS) Quasi-Newton, and Levenberg–Marquardt (LM)—implemented within a backpropagation-based neural network containing a single hidden layer. This setup allowed for efficient and accurate modeling of the experimental results.
Figure 2 illustrates the architecture of the designed (ANN). The input layer contains six neurons, corresponding to the time and force values measured in three independent experiments. The output layer predicts the displacement values. A single hidden layer processes the inputs through weights and biases, which are optimized during the training phase. This configuration—encompassing training and validation steps—was implemented to ensure accurate predictions of the experimental outcomes.
In this study, time and force were used as input variables, while displacement was the predicted output. To ensure effective training and validation, data from the first experiment was randomly split into two subsets: half was used for training the ANN model, and the other half was used for validation. The displacement values from the remaining two experiments were then used as an independent test set, allowing the model to generalize to unseen datasets. This approach ensured that the model was tested on entirely independent data, enhancing its predictive reliability.
o = f ( i w + b ) .
The fundamental principle of the network is given by Equation (4), where o denotes the predicted displacement (output), i the inputs (time and force), w the weight, and b the bias. For each input i in the hidden layer, the network assigns corresponding values of w and b, which are subsequently optimized by the selected training algorithm. During each training epoch, these parameters were adjusted to minimize the error between predicted and actual displacement values—quantified by the Mean Squared Error (MSE)—, thereby enhancing the model’s accuracy.
M S E = ( p i o i ) 2 n .
The performance criterion for stopping iterations is the Mean Squared Error (MSE), set to 10−10, as formulated in Equation (5). Here, p denotes the predicted output, and n represents the number of samples in the dataset [26].

3.1. Scaled Conjugate Gradient

The scaled conjugate gradient (SCG) algorithm differs from other conjugate gradient methods by eliminating the need for a line search at each iteration. This design reduces computational complexity, enhancing efficiency. In the SCG algorithm, the step size is determined through a quadratic approximation of the error criterion, minimizing dependency on user-defined parameters. These attributes make SCG an effective optimization technique for neural network training, particularly in applications requiring faster convergence [27,28].
The step size αk in the SCG algorithm is defined in Equation (6). Here w ¯ represents the weight vector, E ( w ¯ ) is the global error function, p 1 ¯ , p 2 ¯ p k ¯ are non-zero weight vectors, E q w ( w ¯ ) indicates the quadratic approximation of error function, while the μk and δk are parameters that determine the step size.
α k = μ k δ k = p j ¯ T E q w ( y 1 ¯ ) p j ¯ T E q w ( w ¯ ) p j ¯ .
The comparison parameter (Δk) is a scale for how close the quadratic approximation ( E q w ( w ¯ ) ) to the global error function E ( w ¯ + α k p ¯ k ) . If the Δk≥ 0, the error decreases, ensuring convergence [20].
Δ k = 2 δ k E ( w k ¯ ) E ( w k ¯ + α k p k ¯ ) μ k 2 .

3.2. BFGS Quasi-Newton

The BFGS (Broyden Fletcher Goldfarb Shanno) algorithm is a combination of gradient descent and Gauss Newton methods, designed to optimize neural network training. It utilizes a second-order Taylor series approximation for error minimization, as shown in Equation (8). The Hessian matrix (H) is iteratively updated during the process to improve convergence.
H ( x i ) ( x i x ) = f ( x j ) .
The step size is defined based on differences in gradient values between iterations, ensuring stability and efficiency in the optimization process. This step size adjustment, along with the matrix update rule, is represented by Equation (9).
x i x = H ( x j ) 1 f ( x j ) .
The optimization process begins with an identity approximation of the Hessian matrix, expressed as Equation (10). The algorithm then refines the matrix during each iteration, ensuring accurate and efficient convergence through line search methods [22,29,30].
x i x 1 = I . f ( x 0 ) .

3.3. Levenberg—Marquardt

The Levenberg–Marquardt (LM) algorithm addresses non-linear least squares problems by combining Gauss Newton and gradient descent methods [10,12]. This combination improves adaptability and reduces error rates, making LM suitable for high-accuracy neural network training. The gradient descent and Gauss Newton steps are represented in Equations (11) and (12), respectively.
h G D = α J T w y y ^
J T w J h G N = J T w y y ^
A damping parameter (λ) is introduced in Equation (13) to control the optimization process, ensuring stable convergence. Both α in Equation (12) and λ in Equation (13) serve similar roles within the algorithm. The value of λ is crucial and exposes iterative regulations to determine its optimal value. The LM algorithm displays a combined structure that is clarified as
J T w J + λ I . h L M = J T w y y ^ ,
where hLM is a perturbation of the LM. Similarly, in Equation (12), the second-order approximation of Equation (13) is defined by J T w J .
In scenarios involving multiple inputs, as observed in this study, updating the Jacobian matrix (J) at each step becomes compulsory. By applying Broyden’s rank-1 approach for updating J, the renewal process can be expressed as follows:
J n e w = J + y ^ p + h y ^ p J h h T h T h ,
where Jnew is an updated Jacobian matrix of the first derivatives. When comparing the LM algorithm with other algorithms, a key differentiator that makes the LM method stand out is its ability to update the Jacobian matrix (J) dynamically.
The LM algorithm’s ability to dynamically update the Jacobian matrix (J) distinguishes it from other methods. This feature mitigates convergence issues that may arise from using a fixed matrix, ensuring reliable performance across diverse datasets [11].

3.4. Normalization of ANN Parameters

A dataset containing 2700 data points for each input was created based on experimental results. To improve training efficiency, the average values from the experiments were used as the output data. The dataset was normalized using min-max normalization, scaling all data to the range [0, 1], as expressed in Equation (15). This normalization step ensured that all variables were scaled uniformly, reducing the impact of varying magnitudes and facilitating efficient optimization during training.
i n o r m = i i m i n i m a x i m i n
Here, inorm represents the normalized data. Normalization was crucial to ensure consistent network performance and achieve reliable results [31].
The min-max normalization method was selected in this study due to its simplicity and effectiveness in scaling all input variables to a uniform range [0, 1]. This approach ensures that the relative relationships among features are preserved, which is crucial for the stability and convergence of ANN training. Other normalization methods, such as z-score or logarithmic transformations, were not applied as they are more suited to datasets with large variance or Gaussian-distributed features, which were not characteristic of the current dataset. Future work could include comparative analyses of these methods to further validate the advantages of min-max normalization for ANN-based delamination predictions.

4. Results and Discussions

Mode I delamination experiments are needed to calculate the interlaminar delamination toughness. The force-time data will be used to calculate the delamination toughness. Moreover, the force-time data serves as the input to the ANN algorithm, while displacement is the predicted output used for validation.

4.1. Experimental Results

As depicted in Figure 3, the delamination length increases in tandem with the opening displacement of the edges. In Mode I delamination experiments, the applied force similarly rises as the edges are displaced. However, once delamination initiates, the force drops owing to the release of stored energy.
Figure 4 presents the experimental interlaminar delamination toughness values at various opening displacements, alongside the corresponding ANN predictions. At 300 s, for instance, a 10 mm edge opening displacement produces a delamination length of 54 mm (see Figure 3). As the opening displacement grows, the delamination length extends further, reaching approximately 120 mm at a displacement of 40 mm (see Figure 3b). This finding aligns with earlier observations by Roy et al. [32] and Soyugüzel et al. [24], who noted that delamination toughness increases with delamination length after its initiation. In the present study, the average interlaminar delamination toughness is 586.5 J/m² at initiation and 821.7 J/m² during propagation, as illustrated in Figure 4.
The model’s accuracy was evaluated by comparing the predicted displacement values with the actual displacement data obtained from three independent experiments. The average force and time measurements were used as inputs during training to ensure a balanced representation of the experimental conditions. These averaged values were incorporated as training inputs to ensure a balanced representation of the experimental conditions. The error rates for each model were then determined by examining the discrepancies between the predicted outputs and the experimentally derived averages, as illustrated in Figure 5. This approach enabled a robust validation of the network’s ability to generalize and reliably forecast outcomes based on the training data.

4.2. Algorithm Comparison

Evaluating and comparing algorithms is paramount in computational research, as it underpins the selection, optimization, and comprehension of various problem-solving strategies. Such comparative analyses play a pivotal role in advancing computational techniques and technologies. For instance, Arora and Saha [26] compared the LM algorithm with Bayesian Regularization (BR) and Resilient Backpropagation (RP) across seven distinct datasets, identifying strengths and limitations in each approach. Babani et al. [27] compared the SCG and RP algorithms, noting that SCG delivered more stable performance in terms of Mean Squared Error (MSE) and required less training time, while the LM algorithm showed moderate performance relative to these methods. In another study, Mukherjee and Routroy [22] found that the LM algorithm produced fewer errors than the Quasi-Newton algorithm during backpropagation in industrial grinding processes, employing a dataset similar to the one used in the present research.
Performance criteria such as Mean Squared Error (MSE) and true percent error (TPE) generally indicate the overall quality of an artificial neural network. Consequently, the chosen algorithm must meet user-defined thresholds to be deemed suitable. In this section, the performances of the SCG, BFGS, and LM algorithms are compared on the basis of these metrics. The true percentage errors for force and displacement are presented to validate the performance plots
During experimentation, force data exhibited greater incremental changes than time data, while displacement values, used as the predicted output, showed consistent variation across measurements. Even after normalization, the displacement data retain similar magnitudes of variation, which contribute to higher error values, as seen in Table 1 and Table 2. The comparative analysis was conducted using a single hidden layer with 100 neurons, where 80% of the dataset was allocated for training. A cross-validation approach was applied to improve accuracy, ensuring that the test and validation curves operate in parallel and intersect where appropriate, thereby confirming consistency in the training curve.
For all three algorithms, error peaks are most noticeable within the first 200 data points. This tendency can be attributed to fluctuations in experimental measurements, particularly in the displacement dataset. As noted by Karalar et al. [33], the initial segments in experimental data often exhibit higher error values due to the inherent nature of measurement techniques. Consequently, the error at the outset tends to exceed that observed in the middle or latter portions of the dataset, a trend also apparent in the present study.
Figure 6a illustrates the performance of the SCG algorithm on the training set (Test 1), showing that its mean squared error (MSE) reaches a minimum of approximately 10−4 at the 1298th epoch. One notable advantage of SCG is its speed, requiring only 0.0061 s per epoch, which is noticeably faster than the other algorithms tested (Table 1). In contrast, Figure 6b depicts the BFGS algorithm’s results, achieving an MSE of around 10−5 at the 1038th epoch, but requiring 0.0751 s per epoch (Table 1).
Despite the relatively rapid convergence of SCG and the lower MSE of BFGS, additional analysis reveals that the Levenberg–Marquardt (LM) method exhibits more accurate predictions for Test 1. This outcome appears to stem from LM’s ability to fit the underlying force–deformation relationship more precisely when training data points are randomly selected without manual intervention. Meanwhile, in Tests 2 and 3, BFGS demonstrates slightly higher accuracy than LM, although the difference is marginal (See Figure 6).
These observations are further corroborated by Figure 7, Figure 8 and Figure 9, which show that while BFGS and LM deliver comparable performance in the remaining portion of Test 1 (Figure 7), LM’s displacement estimates in Figure 8 and Figure 9 align almost perfectly with the true data points, underscoring LM’s strong predictive capabilities across different segments of the dataset.
Taken together, these results point to LM providing an overall more successful prediction process than BFGS, particularly for Test 1, while BFGS maintains a slight edge in Tests 2 and 3. SCG remains the fastest approach but displays higher errors relative to LM and BFGS.

4.3. Comparison of Hidden Layer Neuron Number

In Section 4.2, the LM method was highlighted for its superior performance, making it a central focus of this study. To further optimize the network’s architecture, a comprehensive evaluation was conducted by varying the number of neurons in a single hidden layer from 10 to 100. The goal was to identify the most efficient configuration based on two primary metrics: percent error (as depicted in Figure 10) and the time required per epoch.
Initial tests using networks with smaller configurations—specifically 10, 20, 30, and 40 neurons—revealed significant limitations. These setups failed to converge effectively and exhibited poor validation performance, resulting in inadequate generalization. Consequently, they were deemed unsuitable and excluded from further analysis. This outcome underscores the importance of selecting an appropriate network size to ensure stability and convergence, particularly when employing the LM method. The remaining configurations, which incorporated larger neuron counts, successfully balanced accuracy and computational efficiency, thereby offering valuable insights into the optimal network architecture for delamination prediction.
As depicted in Table 2, the error observed in networks featuring a hidden layer of 50 neurons remains consistent with those employing higher neuron counts. Specifically, configurations with 60, 70, and even 100 neurons demonstrate a similar error rate of approximately 1%—the highest error recorded for the LM method. This consistency across different configurations is noteworthy, as all tested hidden layers successfully converged and completed 100 validation checks without issue. Nonetheless, when evaluating performance at various hidden layer sizes, the required convergence time emerges as a critical factor. Among the tested configurations, the model with 50 neurons not only maintains high accuracy but also achieves convergence in the shortest total time (00:36), corresponding to 0.0825 s per epoch. This balance between accuracy and computational efficiency underscores the effectiveness of the 50-neuron hidden layer, marking it as the optimal choice for this study.
The selected LM model closely aligns with the experimental results (Figure 7, Figure 8 and Figure 9), underscoring its strong predictive capability for new experimental datasets. This artificial model holds considerable promise for forecasting delamination properties, potentially reducing the need for physical testing on similar specimens. While the LM algorithm’s curve nearly replicates the experimental data, the SCG algorithm exhibits modest deviations. In light of the discussions in Section 4.1 and Section 4.2, the backpropagation-based LM algorithm offers substantial advantages over the other two evaluated methods. Moreover, previous studies by Mukherjee and Routroy [22] and Soepangkat et al. [34] reinforce these findings, consistently demonstrating the LM algorithm’s superior performance.
However, it is worth noting that, despite the LM algorithm’s clear superiority in most metrics, the SCG method—which shows the lowest overall accuracy—can outperform LM in certain aspects, such as faster time per epoch, as also noted by Rosales et al. [35]. Nevertheless, the accuracy of the selected model, reaching approximately 99.5%, underscores its reliability and applicability in predictive delamination studies.
The conclusion that 50 neurons provide the most effective configuration is grounded in a detailed analysis of the trade-off between accuracy and computational efficiency. As illustrated in Table 2, this configuration consistently yielded low error rates and faster convergence compared to other tested neuron counts. However, it is important to note that these experiments were performed under controlled conditions and with a relatively limited dataset.
To further ensure the robustness of this finding, cross-validation was employed using an 80/20 train–test split, repeated over multiple iterations. The results demonstrated that the 100-neuron model sustained stable performance across various runs, exhibiting only minimal fluctuations in error rates. This outcome underscores the consistency of the chosen configuration under the current experimental parameters.
In order to ensure that the dataset accurately represents real-world delamination scenarios, the experimental setup and sample preparation procedures adhered to established ASTM D3171 and D5528 standards, which simulate the most prevalent failure modes observed in aerospace and automotive applications. Although only three unidirectional carbon fiber composite samples were tested, a total of 2700 data points were generated under consistent loading and environmental conditions, providing a diverse yet coherent foundation for training, validation, and testing. This approach enabled an evaluation of the model’s performance on unseen data, revealing robust predictive accuracy and confirming the dataset’s utility in capturing essential force–displacement relationships. Furthermore, the Mode I delamination tests conducted at room temperature and under displacement control reflect common industry practices, bolstering confidence in the practical relevance and generalizability of the results.
While the present dataset and experimental framework proved sufficient for optimizing the network structure in this particular context, future research could broaden the scope by incorporating more diverse datasets and variable experimental conditions. Such expansions would help to validate and refine the optimal neuron count for a wider range of applications.
Figure 10 illustrates the training performance of the Levenberg–Marquardt (LM) method. Across most data points, the error ranges between 0% and 0.10%, demonstrating a high degree of accuracy and robust convergence. Nevertheless, certain points show transient spikes in the error, which could be attributed to outliers or brief mismatches between the model’s predictions and the underlying force–deformation relationships. Overall, the LM algorithm maintains a favorable balance between speed and accuracy under the current experimental conditions. Future work could explore refined data preprocessing and further hyperparameter optimization to mitigate these localized error peaks. These results suggest that the error patterns exhibit a symmetry-like distribution, contributing to the algorithm’s robust convergence and predictive accuracy.

5. Conclusions

This study investigated the effectiveness of three Artificial Neural Network (ANN) algorithms—Scaled Conjugate Gradient (SCG), Broyden–Fletcher–Goldfarb–Shanno (BFGS) Quasi-Newton, and Levenberg–Marquardt (LM)—in predicting Mode-I delamination behavior in composite materials through Double Cantilever Beam (DCB) tests. Among the evaluated algorithms, the LM method achieved the highest accuracy, with prediction rates of 99.6% for force and 99.3% for displacement. Its optimal configuration, featuring 50 neurons in the hidden layer, surpassed both SCG and BFGS in terms of precision and reliability. Although SCG operated at a faster pace, its error rate of 8.6% greatly limited its accuracy. The BFGS algorithm offered a more balanced trade-off between speed and error but remained less precise than LM.
Additionally, this study leveraged a robust dataset of 2700 data points drawn from three separate samples, all prepared and tested under standardized protocols (ASTM D3171, D5528). These conditions effectively capture typical delamination scenarios in aerospace and automotive composites, thereby enhancing the model’s generalization capability. The Levenberg–Marquardt algorithm’s ability to minimize training and validation errors, even with this limited but representative dataset, further underscores its potential for various delamination applications.
The LM algorithm’s robust performance holds considerable implications for industries such as aerospace and automotive, where composite materials are critical for high-strength, low-weight applications. By reducing reliance on extensive physical testing, ANN-driven predictions can significantly cut both development time and resource expenditure, enabling more rapid iterations and optimization of composite structures. Consequently, the LM algorithm’s predictive capabilities present valuable opportunities to enhance and streamline material design processes. The structured patterns in the experimental data contribute to the predictive accuracy and robustness of the ANN models, reflecting symmetry-like properties.
The results highlight the potential of ANNs to minimize experimental efforts in material design by providing accurate and resource-efficient predictions. The proposed framework offers a balanced approach, combining precision and computational efficiency compared to traditional methods. Future work could explore advanced ANN architectures or hybrid AI models to further enhance prediction capabilities and broaden the applicability of this methodology to other failure modes or material systems.
Future investigations should focus on further improving ANN architectures and exploring advanced training methods to enhance both predictive accuracy and computational efficiency. Techniques such as reinforcement learning or genetic algorithms could offer additional optimization benefits. Moreover, extending this approach to predict other failure modes or delamination behaviors would broaden its applicability and provide deeper insights into the performance of composite materials in diverse conditions.
In conclusion, the LM-based ANN model demonstrated high predictive accuracy and practical utility in Mode I delamination prediction, marking a significant step forward in material design optimization. Its capacity to save time and cost makes it a vital asset for material design and testing processes, particularly in sectors where stringent performance standards for composite materials are paramount.
Future research could expand on the findings of this study by exploring alternative network architectures, such as multi-layer or convolutional neural networks, and assessing their performance in conditions involving high noise levels or limited data availability. An additional avenue involves systematically evaluating the robustness of ANN models by introducing controlled noise into the dataset and employing advanced preprocessing techniques—such as smoothing filters, dropout, or weight decay. These directions would provide a broader understanding of ANN capabilities in composite material prediction, enhance their reliability, and further optimize their practical implementation in real-world engineering contexts.

Funding

This research received no external funding.

Data Availability Statement

Data supporting the findings of this study are not currently available as they are part of an ongoing research project. The data will be made available upon completion of the project.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Sha, W.; Guo, Y.; Yuan, Q.; Tang, S.; Zhang, X.; Lu, S.; Guo, X.; Cao, Y.; Cheng, S. Artificial Intelligence to Power the Future of Materials Science and Engineering. Adv. Intell. Syst. 2020, 2, 1900143. [Google Scholar] [CrossRef]
  2. Li, B.; Hou, B.; Yu, W. Applications of artificial intelligence in intelligent manufacturing: A review. Front. Inf. Technol. Electron. Eng. 2017, 18, 86–96. [Google Scholar] [CrossRef]
  3. Liu, Y.; Zhao, T.; Ju, W.; Shi, S. Materials discovery and design using machine learning. J. Materiomics 2017, 3, 159–177. [Google Scholar] [CrossRef]
  4. Vasudevan, R.; Choudhary, K.; Mehta, A.; Smith, R.; Kusne, G.; Tavazza, F.; Hattrick-Simpers, J. Materials science in the artificial intelligence age: High-throughput library generation, machine learning, and a pathway from correlations to the underpinning physics. MRS Commun. 2019, 9, 821–838. [Google Scholar] [CrossRef]
  5. Kang, J.Y.; Choi, B.I.; Lee, H.J. Application of artificial neural network for predicting plain strain fracture toughness using tensile test results. Fatigue Fract. Eng. Mater. Struct. 2006, 29, 321–329. [Google Scholar] [CrossRef]
  6. Guan, K.; Jia, L.; Chen, X.; Weng, J.; Ding, F.; Zhang, H. Improvement of fracture toughness of directionally solidified Nb-silicide in situ composites using artificial neural network. Mater. Sci. Eng. A 2014, 605, 65–72. [Google Scholar] [CrossRef]
  7. Zakaulla, M.; Pasha, Y.; Siddalingappa, S.K. Prediction of mechanical properties for polyetheretherketone composite reinforced with graphene and titanium powder using artificial neural network. Mater. Today Proc. 2022, 49, 1268–1274. [Google Scholar] [CrossRef]
  8. Genel, K. Application of artificial neural network for predicting strain-life fatigue properties of steels on the basis of tensile tests. Int. J. Fatigue 2004, 26, 1027–1035. [Google Scholar] [CrossRef]
  9. Kazi, M.-K.; Eljack, F.; Mahdi, E. Predictive ANN models for varying filler content for cotton fiber/PVC composites based on experimental load-displacement curves. Compos. Struct. 2020, 254, 112885. [Google Scholar] [CrossRef]
  10. Gavin, H.P. The Levenberg-Marquardt Algorithm for Nonlinear Least Squares. Available online: https://people.duke.edu/~hpgavin (accessed on 20 April 2023).
  11. Transtrum, M.K.; Sethna, J.P. Improvements to the Levenberg-Marquardt algorithm for nonlinear least-squares minimization. J. Comput. Phys. 2012, 231, 297–312. [Google Scholar]
  12. Kayri, M. Predictive Abilities of Bayesian Regularization and Levenberg–Marquardt Algorithms in Artificial Neural Networks: A Comparative Empirical Study on Social Data. Math. Comput. Appl. 2016, 21, 20. [Google Scholar] [CrossRef]
  13. Jadhav, P. Modeling of mode 1 delamination behavior in laminated composite structures for aerospace applications. Mater. Today Proc. 2023, 82, 61–67. [Google Scholar] [CrossRef]
  14. Chakraborty, D. Artificial neural network based delamination prediction in laminated composites. Mater. Des. 2005, 26, 1–7. [Google Scholar] [CrossRef]
  15. Khan, A.; Ko, D.-K.; Lim, S.C.; Kim, H.S. Structural vibration-based classification and prediction of delamination in smart composite laminates using deep learning neural network. Compos. Part B Eng. 2019, 161, 586–594. [Google Scholar] [CrossRef]
  16. Sreekanth, T.G.; Senthilkumar, M.; Reddy, S.M. Artificial neural network based delamination prediction in composite plates using vibration signals. Frat. Integrità Strutt. 2022, 63, 37–45. [Google Scholar] [CrossRef]
  17. Bortoluzzi, D.B.; Diniz, C.A.; Pereira, J.L.J.; Ribeiro Junior, R.F.; Gomes, G.F.; Ancelotti Junior, A.C. On the influence of rectangular Z-pins parameters on mode II delamination resistance of through-the-thickness reinforced composites. Compos. Struct. 2023, 305, 116509. [Google Scholar] [CrossRef]
  18. Allegri, G. Modelling fatigue delamination growth in fiber-reinforced composites: Power-law equations or artificial neural networks. Mater. Des. 2018, 155, 59–70. [Google Scholar] [CrossRef]
  19. Sharma, A.; Kumar, S.A.; Kushvaha, V. Effect of aspect ratio on dynamic fracture toughness of particulate polymer composite using artificial neural network. Eng. Fract. Mech. 2020, 228, 106907. [Google Scholar] [CrossRef]
  20. Oliver, G.A.; Ancelotti, A.C., Jr.; Gomes, G.F. Neural network-based damage identification in composite laminated plates using frequency shifts. Neural Comput. Appl. 2021, 33, 3183–3194. [Google Scholar] [CrossRef]
  21. Rautela, M.; Senthilnath, J.; Monaco, E.; Gopalakrishnan, S. Delamination prediction in composite panels using unsupervised-feature learning methods with wavelet-enhanced guided wave representations. arXiv 2022, arXiv:2204.09764. [Google Scholar] [CrossRef]
  22. Mukherjee, I.; Routroy, S. Comparing the performance of neural networks developed by using Levenberg–Marquardt and Quasi-Newton with the gradient descent algorithm for modeling a multiple response grinding process. Expert Syst. Appl. 2012, 39, 2397–2407. [Google Scholar] [CrossRef]
  23. ASTM D3171; Standard Test Methods for Constituent Content of Composite Materials. ASTM International: West Conshohocken, PA, USA, 2015.
  24. Soyugüzel, T.; Mecitoğlu, Z.; Odabası, K.H. Experimental and numerical investigation on the mode I and mode II interlaminar fracture toughness of nitrogen-doped reduced graphene oxide reinforced composites. Theor. Appl. Fract. Mech. 2023, 128, 104103. [Google Scholar] [CrossRef]
  25. ASTM D5528; Mode I Interlaminar Fracture Toughness of Unidirectional Fiber-Reinforced Polymer Matrix Composites. ASTM International: West Conshohocken, PA, USA, 2013.
  26. Arora, I.; Saha, A. Comparison of back propagation training algorithms for software defect prediction. In Proceedings of the 2nd International Conference on Contemporary Computing and Informatics (IC3I), Greater Noida, India, 14–17 December 2016; pp. 51–58. [Google Scholar] [CrossRef]
  27. Babani, L.; Jadhav, S.; Chaudhari, B. Scaled Conjugate Gradient Based Adaptive ANN Control for SVM-DTC Induction Motor Drive. Neurocomputing 2016, 112, 384–395. [Google Scholar] [CrossRef]
  28. Møller, M.F. A Scaled Conjugate Gradient Algorithm for Fast Supervised Learning. Neural Netw. 1993, 6, 525–533. [Google Scholar] [CrossRef]
  29. Kostopoulos, A.E.; Grapsa, T.N. Self-scaled conjugate gradient training algorithms. Neurocomputing 2009, 72, 3000–3019. [Google Scholar] [CrossRef]
  30. Saputro, D.R.S.; Saputro, H.P.; Adiyanto, B. Parameter estimation of Gumbel distribution using Quasi-Newton Broyden Fletcher Goldfarb Shanno (BFGS) method. J. Phys. Conf. Ser. 2019, 1188, 012020. [Google Scholar] [CrossRef]
  31. Mazziotta, M.; Pareto, A. Normalization methods for spatio-temporal analysis of environmental performance: Revisiting the Min–Max method. Environmetrics 2022, 33, e2730. [Google Scholar] [CrossRef]
  32. Roy, S.; Kumar, A. Effect of particle size on mixed-mode fracture of nanographene reinforced epoxy and mode I delamination of its carbon fiber composite. Compos. Struct. 2017, 181, 1–8. [Google Scholar] [CrossRef]
  33. Karalar, A.B.; Soyugüzel, T.; Balkan, D. An Artificial Neural Network Approach to Predict Strain Gauge Results of Unidirectional Laminated Composites’ Tensile Test. In Proceedings of the 2023 10th International Conference on Recent Advances in Air and Space Technologies (RAST), Istanbul, Turkey, 7–9 June 2023; pp. 1–4. [Google Scholar] [CrossRef]
  34. Soepangkat, B.O.P.; Norcahyo, R.; Effendi, M.K.; Pramujati, B. Multi-response optimization of carbon fiber reinforced polymer (CFRP) drilling using back propagation neural network-particle swarm optimization (BPNN-PSO). Eng. Sci. Technol. Int. J. 2020, 23, 700–713. [Google Scholar] [CrossRef]
  35. Rosales, M.A.; Palconit, M.G.B.; Bandala, A.A.; Vicerra, R.R.P.; Dadios, E.P.; Calinao, H. Prediction of Total Body Water using Scaled Conjugate Gradient Artificial Neural Network. In Proceedings of the 2020 IEEE Region 10 Conference (TENCON), Osaka, Japan, 16–19 November 2020; pp. 218–223. [Google Scholar] [CrossRef]
Figure 1. Schematic of Mode I test [24].
Figure 1. Schematic of Mode I test [24].
Symmetry 17 00091 g001
Figure 2. Artificial neural network model.
Figure 2. Artificial neural network model.
Symmetry 17 00091 g002
Figure 3. Mode I delamination experiment captures (a) delamination at 300 s and, (b) delamination at 1200 s.
Figure 3. Mode I delamination experiment captures (a) delamination at 300 s and, (b) delamination at 1200 s.
Symmetry 17 00091 g003
Figure 4. Interlaminar delamination toughness values.
Figure 4. Interlaminar delamination toughness values.
Symmetry 17 00091 g004
Figure 5. Force/Deformation results of mode I delamination experiments.
Figure 5. Force/Deformation results of mode I delamination experiments.
Symmetry 17 00091 g005
Figure 6. Training performances of the algorithms due to MSE, (a) test 1, (b) test 2 and, (c) test 3.
Figure 6. Training performances of the algorithms due to MSE, (a) test 1, (b) test 2 and, (c) test 3.
Symmetry 17 00091 g006
Figure 7. Predicted-Real displacement curves due to force in Test 1, (a) SCG, (b) LM and, (c) BFGS.
Figure 7. Predicted-Real displacement curves due to force in Test 1, (a) SCG, (b) LM and, (c) BFGS.
Symmetry 17 00091 g007
Figure 8. Predicted-Real displacement curves due to force in Test 2, (a) SCG, (b) LM and, (c) BFGS.
Figure 8. Predicted-Real displacement curves due to force in Test 2, (a) SCG, (b) LM and, (c) BFGS.
Symmetry 17 00091 g008
Figure 9. Comparison of the results for the algorithms with 100 neurons, (a) SCG, (b) LM and, (c) BFGS.
Figure 9. Comparison of the results for the algorithms with 100 neurons, (a) SCG, (b) LM and, (c) BFGS.
Symmetry 17 00091 g009
Figure 10. The true percent error in test 1 predictions, Levenberg–Marquardt algorithm.
Figure 10. The true percent error in test 1 predictions, Levenberg–Marquardt algorithm.
Symmetry 17 00091 g010
Table 1. Training time and epochs algorithm comparison.
Table 1. Training time and epochs algorithm comparison.
AlgorithmTotal Time (min:s)EpochsTime/Epoch (s)Maximum Error (Deflection)
SCG00:0812980.00617.20%
BFGS01:1810380.07511.87%
LM00:364360.08250.70%
Table 2. Training results.
Table 2. Training results.
Number of Neurons Total Time (min:s) Epochs Time/Epoch (s) Maximum Error (Deflection)
5000:364360.08250.79%
6000:574690.12310.96%
7001:194890.16330.86%
8002:317210.21000.83%
9002:406070.26360.86%
10002:484930.34230.91%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Balkan, D. Delamination Prediction in Layered Composites Using Optimized ANN Algorithms: A Comparative Analysis. Symmetry 2025, 17, 91. https://doi.org/10.3390/sym17010091

AMA Style

Balkan D. Delamination Prediction in Layered Composites Using Optimized ANN Algorithms: A Comparative Analysis. Symmetry. 2025; 17(1):91. https://doi.org/10.3390/sym17010091

Chicago/Turabian Style

Balkan, Demet. 2025. "Delamination Prediction in Layered Composites Using Optimized ANN Algorithms: A Comparative Analysis" Symmetry 17, no. 1: 91. https://doi.org/10.3390/sym17010091

APA Style

Balkan, D. (2025). Delamination Prediction in Layered Composites Using Optimized ANN Algorithms: A Comparative Analysis. Symmetry, 17(1), 91. https://doi.org/10.3390/sym17010091

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop