PSI - Issue 70

Ch. Vihas et al. / Procedia Structural Integrity 70 (2025) 461–468

466

Model validation is the process of assessing the trained machine learning model's performance on unseen data. It helps us understand whether the model is underfitting or overfitting (Nguyen-Sy, et al (2023)). Figure 3 illustrates the performance of individual models in predicting the compressive strength of concrete (measured in MPa) across three different datasets. Each plot compares the predicted and actual compressive strength of various models. In the training set, all the ML models display blue dots aligned closely to the red dashed ideal line, indicating the predictions are nearly perfect for the training data. For the prediction set, all the ML models exhibit a near-perfect linear relationship with the actual values, demonstrating that the models learned the patterns of the training data very well. In the validation set, the orange dots remain fairly close to the ideal line, although there is a slight spread. This suggests the model performs well on hidden validation data, though there might be slight deviations, possibly due to overfitting or minor inconsistencies in the validation set. Nevertheless, the model appears to generalize well. In the test set, the green dots represent the test data predictions. The pattern remains close to the ideal line, although some outliers exist, particularly for higher compressive strength values. These outliers may indicate challenges the model encounters in predicting extreme values in the test set, but overall, the predictions are quite accurate. The performance metrics were measured for all four machine learning models and are presented in Table 1. From Figure 4, it is observed that XGBoost consistently outperforms the other ML models across training, testing, and validation datasets. It achieves the highest R² value of 0.9751 on validation, indicating strong predictive accuracy, and records the lowest error values in terms of MSE, R², MAE, and RMSE, particularly on the validation set. This suggests that XGBoost fits the training data well and generalizes well on unseen data. Random Forest shows decent performance, with high R² on training (0.9925) but a notable drop on testing and validation, indicating slight overfitting. ANN exhibits moderate performance, with reasonable training metrics but poorer generalization, as seen in the higher RMSE and MAE on testing and especially in the questionable validation RMSE, which may suggest a data inconsistency. Finally, SVM performs the worst among all models, with consistently low R² scores and high error values across all the datasets, and it struggles to identify the patterns in the data. Overall, XGBoost emerges as the most reliable and accurate model for this dataset. 3.2 Performance Metrics

Table 1 . Performance Metrics of Various ML Models

Sl. No.

Machine Learning Model

Dataset

R 2

MSE

RMSE

MAE

1.

Random Forest

Training

0.9925

1.8370

1.3554

0.8143

Testing

0.9163

16.442

4.0549

2.7962

Validation

0.9177

13.256

3.6409

2.6776

2.

XG BOOST

Training

0.9973

0.6668

0.8166

0.6240

Testing

0.9502

9.7794

3.1272

1.9455

Validation

0.9751

4.0164

2.0041

1.4354

3.

ANN

Training

0.9568

10.5923

3.2546

2.2801

Testing

0.9093

17.824

4.4287

3.2667

Validation

0.8783

3.2667

19.6137

4.4287

4.

SVM

Training

0.5303

115.281

10.7369

7.8151

Testing

0.5002

98.2297

9.9111

6.9981

Validation

0.4701

85.363

9.2392

7.6421

Made with FlippingBook - Online catalogs