PSI - Issue 75

Jan Schubnell et al. / Procedia Structural Integrity 75 (2025) 94–101 Schubnell/ Structural Integrity Procedia (2025)

97

4

Table 1. An example of a table. Characteristics of the database on fatigue behavior of engineering steels (Fliegener, J. Rosenberger, M. Luke, J. Domínguez, J. Morgado, H.U. Kobialka, T. Kraft, 2024) Data sources DaBef , Boller&Seeger, NIMS datasheets, JSMS databook, misc. literature see Fliegener et. al. (2024) Scope 22422 data points on specimen level, 1144 on series level, 111 materials Degree of completeness Degree of completeness of the main categories (series with value specified divided by total number of series): Fatigue strength (841 series, 74 %), Hardness HV (1002 series, 88 %), Tensile strength Rm (1007 series, 88 %), Roughness Rz (740 series, 65 %), Stress concentration factor Kt (1124 series, 98 %), Loading type (1110 series, 97 %), R ratio (1091 series, 95 %), cross section geometry (845, 74 %)

Steel groups

According to DIN 10027-2, JIS, AISI, SAE: - Alloyed construction, engineering and pressure vessel steels - Construction steels - Heat-resisting and stainless steels - Low-alloyed construction, engineering and pressure vessel steels - Other steels - Roller bearing steels - Spring steels - Tool steels - Unalloyed construction, engineering and pressure vessel steels

Table 2. Data spaces used for ML training Data space No. of SN curves No. of specimen

Materials

Description

DS1 DS2 DS3 DS4

1152 1152

111 materials 111 materials 49 materials 49 materials

22527 12030

Include all data

No runouts, fatigue life < 10 7 No runouts, fatigue life < 2 × 10 6 , < 1000 MPa No runouts, fatigue life < 2 × 10 6 , < 1000 MPa, specimen averaged according to equation (2) for specimen at equal No runouts, 10 4 < fatigue life < 2 × 10 6 , < 1000 MPa, specimen averaged according to equation (2) for specimen at equal

477 305

8892 3092

DS5 DS6

305 109

2709 1095

49 materials

100Cr6

4. Application of ML approach Two different machine learning approaches was used in this work, illustrated in Figure 2. First, the decision tree based Random Forrest approach was used according to former work (Fliegener, J. Rosenberger, M. Luke, J. Domínguez, J. Morgado, H.U. Kobialka, T. Kraft, 2024), illustrated for an hardness prediction in Figure 2 (a). Second, a shall artificial neural network is used, shown in Figure 2 (b). A random forest regressor is a machine learning algorithm that falls under the ensemble learning category, specifically within the bagging methods family (Breiman, 2001). For each model estimator, a decision tree is grown. Each tree is trained on a distinct bootstrapped subset of the training data to enhance diversity among the decision trees. This amplifies the overall generalizability of the random forest model and reduces the risk of overfitting. In the bootstrapping method (Feron, 1979), the fundamental assumption is that the random sample is "representative" of the population it was drawn from. This population is now substituted by the sample. New samples are generated by repeated sampling with replacement. When splitting each node during the tree's construction, the best split is identified through an exhaustive search of the values of either all input features or a random subset of size max features. According to (Pedregosa, 2011), considering all features is an empirically sound default for regression tasks. During prediction, each tree in the forest yields an output. For regression tasks, the final prediction is typically the average of the outputs from all the trees.

Made with FlippingBook flipbook maker